SlideShare une entreprise Scribd logo
1  sur  26
Télécharger pour lire hors ligne
Purdue University
Purdue e-Pubs
ECE Technical Reports Electrical and Computer Engineering
9-1-1992
Implementation of back-propagation neural
networks with MatLab
Jamshid Nazari
Purdue University School of Electrical Engineering
Okan K. Ersoy
Purdue University School of Electrical Engineering
This document has been made available through Purdue e-Pubs, a service of the Purdue University Libraries. Please contact epubs@purdue.edu for
additional information.
Nazari, Jamshid and Ersoy, Okan K., "Implementation of back-propagation neural networks with MatLab" (1992). ECE Technical
Reports. Paper 275.
http://docs.lib.purdue.edu/ecetr/275
TR-EE 92-39
SEF~EMBER1992
TABLE OF CONTENTS
Page
LIST OF TABLES . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . v
LIST OF FIGURES . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . vi
ABSTRACT . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . vii
1. BACK PROPAGATION ALGORITHM USING MATLAB . . . . . . . . 1
What is Matlab? . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
Why Use Matlab? . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
Speed Comparison of Matrix Multiply in Matlab and C . . . . . . . 2
Back Propagation Algorithm . . . . . . . . . . . . . . . . . . . . . . . 3
Mbackprop Program . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
Reducing Number of Iterations Increases Execution Speed . . . . . 4
Speed Comparison of Algorithm 1and Algorithm 2 . . . . . . . . . 5
Matlab Backprop Speed vs. C Backprop Speed . . . . . . . . . . . . 7
Integrated Graphical Capability of the Mbackprop Program . . . . . 9
Other Capabilities of the Mbackprop Package . . . . . . . . . . . . . . 10
Summary and Conclusions . . . . . . . . . . . . . . . . . . . . . . . . 11
. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .BIBLIOGRAPHY I!,
LIST OF TABLES
Page
Speed Comparison of Matrix Multiply in Matlab and a C program . . 2
Algorithm 1 Solves Class Identification Problem . . . . . . . . . . . . 6
. . . . . . . . . . . .Algorithm 2 Solves Class Identification Problem 7
. . . . . . . . . . . . . . .Size of the Variables in Algorithms1 and 2 8
Speed of Algorithm1 vs. Speed of Algorithm 2 . . . . . . . . . . . . 9
Speed of Matlab Backprop Program vs. Speed of C Backprop Programs 9
. . .Comparison of Speeds in Single and Double Precision Backprops 10
LIST OF FIGURES
Figure
1.1
1.2
1.3
1.4
1.5
1.6
1.7
I't~gt-
. . . . . . . . . . . . . . . . . . . . . .Variables Used in Algorithm 1 12
. . . . . . . . . . . . . . . . . . . . . .Variables Used in Algorithm 2 13
Sample Mean Square Error Graph Generated by Mbackprop . . . . . 14
. . . . . . .Sample Percent Correct Graph Generated by Mbackprop 15
Sample Maximum Absolute Error Graph Generated by Mbackprop . . 16
. . . . .Sample Percent Bits Wrong Graph Generated by Mbackprop 17
. . . . . . . . . . .Sample Compact Graph Generated by Mbackprop 18
ABSTRACT
The artificial neural network back propagation algorithm is implemented in Mat-
lab language. This implementation is compared with severalother software packages.
The effect of reducing the number of iterations in the performance of the algorithm
iai studied. The speed of the back propagation program, mkckpmp, written in Mat-
lab language is compared with the speed of several other back propagation programs
which are written in the C language. The speed of the Matlab program mbackpmp
is,also compared with the C program quickpmp which is a variant of the back prop-
agation algorithm. It is shown that the Matlab program mbackpmp is about 4.5 to 7
times faster than the C programs.
1. BACK PROPAGATION ALGORITHM USING MATLAB
This chapter explains the software package, mbackprop, which is written in MatJah
language. The package implements the Back Propagation (BP) algorithm [RIIW861,
which is an artificial neural network algorithm.
There are other software packages which implement the back propagation algo-
rithm. For example the AspirinIMIGRAINESSoftware Tools [Leig'I]is intended to be
used to investigate different neural network paradigms. There is also NASA NETS
[Baf89] which is a neural network simulator. It provides a system for a variety of
neural network configurations which uses generalized delta back propagation learn-
ing method. There are also books which have implementation of BP algorithm in C
language for example, see [ED90].
Many of these software packages are huge, they need to be compiled and some-
times difficult to understand. Modification of these codes requires understanding the
rnassive amount of source code and additional low level programming. The mbackprop
on the other hand is easy to use and very fast. With the graphical capability of the
Idatlab the network parameters can be graphed to see what is going on inside any
specific network. Additions and modifications to the mbackprop package are easier
a~ndfurther research in the area of neural network can be facilitated.
1.1 What is Matlab?
Matlab is a commercial software developed by Mathworks Inc. It is an interactive
software package for scientific and engineering numeric computation [Inc90]. Matlab
has several basic routines which do matrix arithmetics, plotting etc.
1..2 Why Use Matlab?
Matlab is already in use in many institutions. It is used in research in academia
and industry. Prototype solutions are usually obtained faster in Matlab than solving
a, problem from a programming language.
Matlab is fast, because the core routines in Matlab are fine tuned for diflerent
computer architectures. Following test was made to compare the speed between
Matlab and a program written in C. Since the back propagation algorithm involves
nnatrix manipulations the test chosen was matrix multiply. As the next section shows,
ndatlab1 was about 2.5 times faster than a C program both doing a matrix multiply.
!?:peedComparison of Matrix Multiply in Matlab and C
A program in C was written to multiply two matrices containing double precision
numbers. The result of the multiplication is assigned into a third matrix. Each
matrix contained 500 rows and 500 columns. A Matlab M file was written to do
the same multiply as C program did. Only the segment of the code which does the
nlultiplication is timed. The test was run on an IPC-SparcStation computer, the
rlesult is shown in Table 1.1. As the table shows Matlab is faster than the C program
bly more than a factor of two.
Table 1.1 Speed comparison of matrix multiply in Matlab and a C program. Matlab
runs 2.5 times faster than the C program.
'The version of Matlab we used waa 3.5i.
1..3 Back Propagation Algorithm
The generalized delta rule [RHWSG], also known as back propagation algorit,li~n
is explained here briefly for feed forward Neural Network (NN). The explanitt,ion Ilcrc
is intended to give an outline of the process involved in back propagation algorithm.
The NN explained here contains three layers. These are input, hidden, and output
Layers. During the training phase, the training data is fed into to the input layer. The
dlata is propagated to the hidden layer and then to the output layer. This is called
the forward pass of the back propagation algorithm. In forward pass, each node in
hidden layer gets input from all the nodes from input layer, which are multiplied with
appropriate weights and then summed. The output of the hidden node is the non-
linear transformation of the this resulting sum. Similarly each node in output layer
gets input from all the nodes from hidden layer, which are multiplied with appropriate
weights and then summed. The output of this node is the non-linear transformation
of the resulting sum.
The output values of the output layer are compared with the target output values.
The target output values are those that we attempt to teach our network. The error
between actual output values and target output values is calculated and propagated
back toward hidden layer. This is called the backward pass of the back propagation
algorithm. The error is used to update the connection strengths between nodes, i.e.
weight matrices between input-hidden layers and hidden-output layers are updated.
During the testing phase, no learning takes place i.e., weight matrices are not
changed. Each test vector is fed into the input layer. The feed forward of the testing
data is similar to the feed forward of the training data.
1..4 Mbackprop Program
The mbackprop program is written in Matlab language. The program implements
the back propagation algorithm [RHW86]. The algorithms used in the nabackprop
program involve very few number of iterations. This is one of the reasons why this
program is so fast. In the next section, an example is given to see the effect of reducing
number of iterations has on the executionspeed of a program. In Section1.5execution
speed of the mbackprop program in Matlab is compared with the execution speed of
a back propagation program in C.
fiducing Number of Iterations Increases Execution Spccd
There are several ways to write a program to accomplish a given task. The
approach or algorithm a person might take will have a great effect on the execution
s p d of a program. Here, a class identification problem is stated and then two
solutions are presented. Statement of the problem is, given a matrix A, find the class
to which each column of the matrix A belongs.
Each column of the matrix A is a vector x which we want tofind to which class this
veztor belongs. To do this, for each of these vectors x, we want to find the distances
between the vector x and m other vectors. These m vectors are the desired vectors
representing classl through class,. The minimum of the m distances, comes from a
vector representing class,. The number j is the answer to the column vector x. So
the desired output is a row vector B indicating to which class each of the vectors in A
belongs. Content of the matrix A is changing, so we need to calculate the row vector
B more than once.
Two solutions are now presented for the above problem. The first solution will be
algorithm 1 and the second solution will be algorithm 2. Both algorithms will need
az~input argument the following variables:
variable "A" which contains the matrix A
variable "Classes" which contains vectors representing classr through class,
variable "nClassesn which contains the number of classes rn.
The output of the both algorithms is variable "B" which will contain the class number
of each column of the variable "A".
Figure1.1shows severalof the variablesused in algorithm 1. Here the variable "An
is made of columns XI, 2 2 , ...,x,. Variable 'Classes" is made of columns cl,c2,...,cm
which represents classl through class,. Variable 'dist" is a column vector of size m
which will hold the distance of a vector x in A to each of the m classes in variable
"Classesn
. The algorithm 1 is the following:
for each xi in A where i = 1,...,n
- dist(j) = Square Euclidean Distance( x;, c; ) where j = 1,...,m
- B(i) = k where dist(k) = min( dist )
Figure 1.2 shows several of the variables used in algorithm 2. Here the variable
'An
is also made of columns xl,x2,...,x, but we will view it as one block. Variable
'Classesn
is made of m column blocks e l , ...,C,, where m is the number of classes.
Each block Cj is the same size as block A. The block C; contains n equal columns
where n is the number of columns in A. Each column in block Cj is cj which represents
classj. Variable "distn
is made of m block columns which will hold the distance of
bllock A to each of the m blocks in variable uClasses". The algorithm 2 is the following:
dist(j,:) = Square Euclidean Distance( A, Cj ) where j = l!,...,m. dist(j,:)
refers to row j of dist matrix.
B(i) = k where dist(k,i) = min( did(:, i) ). did(:, i) refers to column i of dist
matrix.
Tables 1.2 and 1.3 show the two solutions for the class identification problem
using algorithms 1 and 2. Note that these solutions are written in 'Matlab language.
The algorithm 1 used in Table 1.2 is straight forward. As shown in the next section,
the algorithm 1 contains much more iterations than algorithm 2. This causing the
aJgorithm 1 to run slower than the algorithm 2 of Table 1.3.
Speed Comparison of Algorithm 1 and Algorithm 2
The above algorithms were used to solve the class identification problem, where
the number of classes was 8. The size of the variables used in algorithms 1 and 2
Table 1.2 Algorithm 1 is a straight forward method which solves the class identifica-
tion problem.
1) function B=algorithml( A, Clseeerr, nClasses )
2) % Each column of the variable "Clesseenrepresents a class
3) [ n b w , nCol ] = size( A );
4) B = sem(1, nCol ); % Preallocate memory
5) for i = 1: nCol,
6) x = A ( : , i ) ;
7) for j = 1 : nclaeses,
8) dist( j ) = sum( ( x - Claeses( :,j ) ) .- 2 );
9) end
10) [ v, B(i) ] = min( dist );
11) end
are shown in Table 1.4. Note that the amount of memory used by algorithm 2 (1922
Kbytes) is muchgreater than the memory used in algorithm 1 (212 Kbytes). However,
as shown below, algorithm 2 is much faster than algorithm 1. The speed of execution
is related to the number of iterations in the algorithm.
The number of iterations for algorithm 1 is much greater than the number of
iterations for algorithm 2. In this example, the statement number 8 in algorithm 1
gets executed 24,000 (nCol x nClasses = 3000 x 8) times. Where in algorithm 2
either statement number 11 or 15 gets executed only 8 (nclasses = 8) times. Since
hdatlab is an interpretive language algorithm 1 is much slower than algorithm 2.
Table 1.5 shows that algorithm 2 runs about 23 times faster than algorithm 1. The
test waa performed on an IPC-SparcStation computer. In the next section the speed
of mbackpmp program, written in Matlab, is compared to the speed of a C program
bloth implementing the back propagation algorithm [RHW861.
Table 1.3 Algorithm 2 is another way to solve the class identification problem. It is
:faster than Algorithm 1.
1) function B=algorithm2( A, Classes, nClasses )
2) % Each column of the variable "Classesn
represents all of the
3) % "nclasses" classes. If there are 8 clasaea and each class is
4) % represented by 8 numbers, then the number of rows of "Classes" is
5) % equal to 64. The number of columns in "Classes" is equal to number
6) % of columns in A.
7) [ n b w , nCol ] = size( A );
8) dist = zeros( nclasses, nCol ); % Preallocate memory
9) if n b w == 1
10) for j = 1 : nclasses,
11) dist( j, : ) = ( A - Classes( j, : ) ) .' 2;
12) end
13) else
14) forj=l:nClasses,
15) dist(j,:)=sum((A-Classes(((j-l)*nRow+l):(j*nRow),:)) . - 2 ) ;
16) end
17) end
18) [ v, B ] = min( dist );
yll.5 Matlab Backprop Speed vs. C Backprop Speed
The back propagation program in Matlab, mbackprop, is compared with two other
(Z back propagation programs fbackprop2
and dbackprop. The mbackprop is also com-
pared with the C program quickprop [Fah88]. The quickprop program is a modification
of a back propagation program which has similar feed forward and backward routines
but in update weight routine, all the weights are updated as a function of each weight's
current slope, previous slope, and the size of the last jump. However, if the variable
"ModeSwitchThreshold" in the quickprop program is set to a big number then all the
weight updates are based on normal gradient descent method i.e. same as in regular
back propagation algorithm.
The program fbackprop is similar to the program dbackprop. The only differenceis
that the calculations in fbackpmpare in floats (single precision), where the calculations
aThe jbaekprop program has been used in some of our research at Purdue University for last two
years.
'I'able 1.4 Size of the variables in algorithms 1 and 2 is shown here. The amount of
memory used in algorithm 1 is 212 Kbytes where algorithm 2 uses 1922 Kbytcs.
in dbackpropare in doubles (double precision). All the calculations in Matlab program
Variable
Name
A
B
Clasees
dist
I
j
nClasea
nCol
n h w
v
x
tirzbackpropare in doubles. The calculations in the quickprop program are in floats.
In the fbackprop and the dbackprop programs, weights get updated after every in-
Siee of Variable
in Algorithm 1
8 x 3000
1 x 3000
8 x 8
8 x 1
1 x 1
1 x 1
1 x 1
1 x 1
1 x 1
1 x 1
8 x 1
p,ut/output vector pair. Where the weights in quickpmp and mbackprop programs get
updated after a complete sweep of the training data. As shown below the mbackprop
Siw of Variable
in Algorithm 2
8 x 3000
1 x 3000
64 x 3000
8 x 3000
1 x 1
1 x 1
1 x 1
1 x 1
1 x 3000
program is faster than all the three C programs.
The neural network, used in our benchmark tests, had 64 input nodes, 16 hidden
#Bytea
Alg 1
192,000
24,000
512
64
8
8
8
8
8
8
64
nodes, and 8 output nodes. The training data contained 1600 input and output vector
pairs. Each input vector was 64 numbers and each output vector was 8 numbers. The
#Bytes
Alg 2
192,000
24,000
1,536,000
192,000
8
8
8
8
24,000
training timefor 100 sweepsover the training data is measured for the above programs
using an IPC-SparcStation Computer. The results are shown in Table 1.6. As the
tableshows, the mbackprop runs 7.0 times faster than the C program dbackprop3
. The
-
mbackprop runs 4.5 times faster than the C program quickprop.
The training time for the C programs j5ackprop4 and dbackprop is also measured
in two other computerss. One of the computers was a Vax 11/780 and the other
3The training time for jhockpmp, dbockpmp, and quickpmp are measured for 10 iterations. To
get the training time for 100 iterations, the measured nurnbcrcl are multiplied by 10.
h he training time for fbackpmp and dbackpmp are measured for 1 iteration. To get the training
time for 100 iterations, the measured numbers are multiplied by 100.
6Wedid not have Matlab running in thee computers
Table 1.5 Speed of algorithm 1 is compared to the speed of algorithm 2. Algorithm
2 runs about 23 times faster than algorithm 1.
Table 1.6 Speed of the Matlab program mbackprop is compared to the speed of C
programs fbackprop, dbackprop and quickprop. The training time for 100 swcvps ovcr
the training data is measured. The Matlab program mbackprop runs 7.0 times faster
than the C program dbackprop. The mbackprop also runs 4.5 times faster than the
quickprop program.
Program Execution Time
fbackprop 3969 seconds
dbackprop 3792 seconds
mbackprop 536 seconds
quickprop 2407 secondsi
computer was a Zenith 386133 running SCO unix with math coprocessor. Table 1.7
ishows the training time for 100 iterations over the training data. The time taken for
the fbackpropprogram was less than the dbackprop program in these computers, how-
ever the fbackprop time of the IPC-SparcStation computer was longer the dbackprop
time. So depending on different computer architectures floating point single precision
calculations are faster than double precision calculations or vice versa.
As it was shown, the mbackprop program was fastest among the programs we
considered above. However mbackprop provides an integrated graphic capability that
other programs lack.
1..6 Integrated Graphical Capability of the Mbackprop Program
The back propagation program mbackprop is faster than the C programs consid-
ered here. However this is not the only advantage that this program has over others.
It provides an integrated graphical capability and an interprative environment that
other programs lack.
Table 1.7 Execution speed of the fbackprop, a single precision back propagation pro-
gram, is compared to the dbackprop a double precision back propagation prograrrr.
The training time for 100 sweeps over the training data is nlca..urc:d o ~ lI,hru: co~rl-
puters. In the IPC-SparcStation computer double precision program was fmtc?rthan
t'he single precision program. In Vax 11/780 and 386 computer the single precision
program is faster than the double precision program. The execrltion time of the
rnbackprop is included here for comparison.
In the mbackprop program, the network parameters can be easily viewed during
program execution. Training and testing reports can be enabled during training and
statistics such as mean square error, percent correct, etc. can be collected in report
intervals specified by the user.
Figure 1.3 shows a sample 'mean square error'graph. Figure 1.4 shows a sample
'percent correct'graph. Percent correct refers to the percentage of the input vectors,
in training or testing data, which are correctly classified. Figure 1.5 shows a sample
'mazimum absolute e m r ' graph. Figure 1.6 shows a sample 'percent bits wrong'
graph. Percent bits wrong refers to the percentage of the output nodes which were
off more than some threshold. Figure 1.7 shows a sample 'compact graph' graph.
Compact graph is a graph which contains the above 4 graphs in one graph.
Other graphs can be easily added to the mbackprop package. In the next section
vie list several other capabilities of the mbackprop program.
Computer
IPGSparcStation
Vax 11/780
Zenith 386/33
1.7 Other Capabilities of the Mbackprop Package
So far we haveshown that mbackproppackage is fast and contains several standard
graphical capabilities. Several of the mbackprop capabilities are:
Execution Time
for fbaekprop
3,969 seconds
69,923 seconds
21,795seconds
a allows a user to specify a weight file for initial weights to start training.
Execution Time
for Dbackprop
3,792 seconds
73,150 seconds
24,523seconds
Execution Time
for Mbackprop
536 seconds
can generate random initial weights for training and allows the user to save
these initial weights to be used later.
if the training gets started in wrong initial weights, the program is easily inter-
rupted and different set of initial weights is used.
the result from training a network can be saved and recalled at a later time.
allows further training from where it was last left off.
l.8 Summary and Conclusions
The mbackprop program is written in Matlab language. This program implements
loheback propagation algorithm [RHW86]. Since Matlab is an interpretive language
the number of iterations in the algorithms have been reduced. This reduced number
of iterations results in a faster executable program. The mbackprop program is faster
than the C back propagation program dbackprop by a factor of 7.0. It is faster
than quickprop [Fah88] program by a factor of 4.5. The mbackprop provides other
capabilities such as integrated graphics and interpretive environment which Matlab
offers.
The mbackprop size is less than a comparative program in C. It is modular and each
individual module can be viewed as a software Integrated Chip (IC). Each software
IC can be modified as long as the input/output criteria is met. Additions of other
software ICs is easy to be incorporated into the mbackprop package. Further research
in the area of neural network can be facilitated.
Figure 1.1 Some of the variables used in algorithm 1 is shown here.
Classes=
dist =
Figure 1.2 Some of the variables used in algorithm 2 is shown here.
Figure1.3 A sample 'mean square emr'graph, generated by the mbackpropprogram,
it3 shown here. The solid line is the training mean square error and the dashed line is
the testing mean square error.
Figure 1.4 A sample 'percent correct'graph, generated by the mbnckprop program,
is shown here. The solid line is training percent correct and the dashed line is the
testing percent correct.
Figure 1.5 A sample 'maximum absolute error'graph, generated by the mbackprop
program, is shown here. The solid line is training maximum absolute error and the
dashed line is the testing maximumabsolute error.
Figure 1.6 A sample 'percent bits wrong'graph, generated by the mbnckpropprogram,
is shown here, The solid line is the training percent bits wrong and the dashed line
is; the testing percent bits wrong.
0 ' I I
0 50 100
IterU
0' I I
0 50 100
IterU
Figure 1.7 A sample 'compact 'graph, generated by the mbackprop program, is shown
here. The solid lines refers to the training results and the dashed lines refer to the
testing results.
BIBLIOGRAPHY
BIBLIOGRAPHY
[Baf89] Paul T. Baffes. NETS Users's Guide, Version 2.0 of NETS. Technical Re-
port JSC-23366, NASA, Software Technology Branch, Lyndon B. Johnson
Space Center, September 1989.
[ED901 R. C. Eberhart and R. W. Dobbins. Neuml Network PC Tools, A Practical
Guide. Academic Press, San Diego, California 92101, 1990.
[Fah88] E. Fahlman, Scott. An Empirical Study of Learning Speed in Back-
Propagation Networks. Technical Report CMU-CS-88-162, CMU, CMU,
September 1988.
[Inc9OJ The MathWorks Inc. PRO-MATLAB for Sun Workstations, User's Guide.
The MathWorks Inc., January 1990.
[Lei911 Russell R. Leighton. The Aspirin/MIGRAINES Software Tools, User's
Manual, Release V5.0. Technical Report MP-91W00050, MITRE Corpo-
ration, MITRE Corporation, December 1991.
[RHW86] D. E. Rumelhart, G. E. Hinton, and R. J. Williams. Leaning Internal
Representaions by Ewvr Propagation in Rumelhart, D. E. and McClelland,
J. L., Pamllel Distributed Processing: Explorations in the Microstructure
of Cognition. MIT Press, Cambridge Massachusette, 1986.

Contenu connexe

Tendances

Comprehensive Performance Evaluation on Multiplication of Matrices using MPI
Comprehensive Performance Evaluation on Multiplication of Matrices using MPIComprehensive Performance Evaluation on Multiplication of Matrices using MPI
Comprehensive Performance Evaluation on Multiplication of Matrices using MPIijtsrd
 
Large data with Scikit-learn - Boston Data Mining Meetup - Alex Perrier
Large data with Scikit-learn - Boston Data Mining Meetup  - Alex PerrierLarge data with Scikit-learn - Boston Data Mining Meetup  - Alex Perrier
Large data with Scikit-learn - Boston Data Mining Meetup - Alex PerrierAlexis Perrier
 
Elementary Parallel Algorithms
Elementary Parallel AlgorithmsElementary Parallel Algorithms
Elementary Parallel AlgorithmsHeman Pathak
 
Fpga based efficient multiplier for image processing applications using recur...
Fpga based efficient multiplier for image processing applications using recur...Fpga based efficient multiplier for image processing applications using recur...
Fpga based efficient multiplier for image processing applications using recur...VLSICS Design
 
DESIGN OF DELAY COMPUTATION METHOD FOR CYCLOTOMIC FAST FOURIER TRANSFORM
DESIGN OF DELAY COMPUTATION METHOD FOR CYCLOTOMIC FAST FOURIER TRANSFORMDESIGN OF DELAY COMPUTATION METHOD FOR CYCLOTOMIC FAST FOURIER TRANSFORM
DESIGN OF DELAY COMPUTATION METHOD FOR CYCLOTOMIC FAST FOURIER TRANSFORMsipij
 
Parallel Algorithms
Parallel AlgorithmsParallel Algorithms
Parallel AlgorithmsHeman Pathak
 
Rapport_Cemracs2012
Rapport_Cemracs2012Rapport_Cemracs2012
Rapport_Cemracs2012Jussara F.M.
 
PRAM algorithms from deepika
PRAM algorithms from deepikaPRAM algorithms from deepika
PRAM algorithms from deepikaguest1f4fb3
 
GRAPH MATCHING ALGORITHM FOR TASK ASSIGNMENT PROBLEM
GRAPH MATCHING ALGORITHM FOR TASK ASSIGNMENT PROBLEMGRAPH MATCHING ALGORITHM FOR TASK ASSIGNMENT PROBLEM
GRAPH MATCHING ALGORITHM FOR TASK ASSIGNMENT PROBLEMIJCSEA Journal
 
Optimizing an Earth Science Atmospheric Application with the OmpSs Programmin...
Optimizing an Earth Science Atmospheric Application with the OmpSs Programmin...Optimizing an Earth Science Atmospheric Application with the OmpSs Programmin...
Optimizing an Earth Science Atmospheric Application with the OmpSs Programmin...George Markomanolis
 
Parallel algorithms
Parallel algorithmsParallel algorithms
Parallel algorithmsguest084d20
 
Pretzel: optimized Machine Learning framework for low-latency and high throug...
Pretzel: optimized Machine Learning framework for low-latency and high throug...Pretzel: optimized Machine Learning framework for low-latency and high throug...
Pretzel: optimized Machine Learning framework for low-latency and high throug...NECST Lab @ Politecnico di Milano
 
PRETZEL: Opening the Black Box of Machine Learning Prediction Serving Systems
PRETZEL: Opening the Black Box of Machine Learning Prediction Serving SystemsPRETZEL: Opening the Black Box of Machine Learning Prediction Serving Systems
PRETZEL: Opening the Black Box of Machine Learning Prediction Serving SystemsNECST Lab @ Politecnico di Milano
 
Collective Communications in MPI
 Collective Communications in MPI Collective Communications in MPI
Collective Communications in MPIHanif Durad
 
GPUFish_technical_report
GPUFish_technical_reportGPUFish_technical_report
GPUFish_technical_reportCharles Hubbard
 
Hybrid Model Based Testing Tool Architecture for Exascale Computing System
Hybrid Model Based Testing Tool Architecture for Exascale Computing SystemHybrid Model Based Testing Tool Architecture for Exascale Computing System
Hybrid Model Based Testing Tool Architecture for Exascale Computing SystemCSCJournals
 

Tendances (20)

Comprehensive Performance Evaluation on Multiplication of Matrices using MPI
Comprehensive Performance Evaluation on Multiplication of Matrices using MPIComprehensive Performance Evaluation on Multiplication of Matrices using MPI
Comprehensive Performance Evaluation on Multiplication of Matrices using MPI
 
Large data with Scikit-learn - Boston Data Mining Meetup - Alex Perrier
Large data with Scikit-learn - Boston Data Mining Meetup  - Alex PerrierLarge data with Scikit-learn - Boston Data Mining Meetup  - Alex Perrier
Large data with Scikit-learn - Boston Data Mining Meetup - Alex Perrier
 
Matopt
MatoptMatopt
Matopt
 
Elementary Parallel Algorithms
Elementary Parallel AlgorithmsElementary Parallel Algorithms
Elementary Parallel Algorithms
 
Fpga based efficient multiplier for image processing applications using recur...
Fpga based efficient multiplier for image processing applications using recur...Fpga based efficient multiplier for image processing applications using recur...
Fpga based efficient multiplier for image processing applications using recur...
 
DESIGN OF DELAY COMPUTATION METHOD FOR CYCLOTOMIC FAST FOURIER TRANSFORM
DESIGN OF DELAY COMPUTATION METHOD FOR CYCLOTOMIC FAST FOURIER TRANSFORMDESIGN OF DELAY COMPUTATION METHOD FOR CYCLOTOMIC FAST FOURIER TRANSFORM
DESIGN OF DELAY COMPUTATION METHOD FOR CYCLOTOMIC FAST FOURIER TRANSFORM
 
Parallel Algorithms
Parallel AlgorithmsParallel Algorithms
Parallel Algorithms
 
Rapport_Cemracs2012
Rapport_Cemracs2012Rapport_Cemracs2012
Rapport_Cemracs2012
 
PRAM algorithms from deepika
PRAM algorithms from deepikaPRAM algorithms from deepika
PRAM algorithms from deepika
 
GRAPH MATCHING ALGORITHM FOR TASK ASSIGNMENT PROBLEM
GRAPH MATCHING ALGORITHM FOR TASK ASSIGNMENT PROBLEMGRAPH MATCHING ALGORITHM FOR TASK ASSIGNMENT PROBLEM
GRAPH MATCHING ALGORITHM FOR TASK ASSIGNMENT PROBLEM
 
Optimizing an Earth Science Atmospheric Application with the OmpSs Programmin...
Optimizing an Earth Science Atmospheric Application with the OmpSs Programmin...Optimizing an Earth Science Atmospheric Application with the OmpSs Programmin...
Optimizing an Earth Science Atmospheric Application with the OmpSs Programmin...
 
Parallel algorithms
Parallel algorithmsParallel algorithms
Parallel algorithms
 
Pretzel: optimized Machine Learning framework for low-latency and high throug...
Pretzel: optimized Machine Learning framework for low-latency and high throug...Pretzel: optimized Machine Learning framework for low-latency and high throug...
Pretzel: optimized Machine Learning framework for low-latency and high throug...
 
SparkNet presentation
SparkNet presentationSparkNet presentation
SparkNet presentation
 
PRETZEL: Opening the Black Box of Machine Learning Prediction Serving Systems
PRETZEL: Opening the Black Box of Machine Learning Prediction Serving SystemsPRETZEL: Opening the Black Box of Machine Learning Prediction Serving Systems
PRETZEL: Opening the Black Box of Machine Learning Prediction Serving Systems
 
Chapter 4 pc
Chapter 4 pcChapter 4 pc
Chapter 4 pc
 
Collective Communications in MPI
 Collective Communications in MPI Collective Communications in MPI
Collective Communications in MPI
 
All About MATLAB
All About MATLABAll About MATLAB
All About MATLAB
 
GPUFish_technical_report
GPUFish_technical_reportGPUFish_technical_report
GPUFish_technical_report
 
Hybrid Model Based Testing Tool Architecture for Exascale Computing System
Hybrid Model Based Testing Tool Architecture for Exascale Computing SystemHybrid Model Based Testing Tool Architecture for Exascale Computing System
Hybrid Model Based Testing Tool Architecture for Exascale Computing System
 

En vedette

Derpless Discourse: authentic promotion for responding to Forrester, Gartner,...
Derpless Discourse: authentic promotion for responding to Forrester, Gartner,...Derpless Discourse: authentic promotion for responding to Forrester, Gartner,...
Derpless Discourse: authentic promotion for responding to Forrester, Gartner,...Glen Ford
 
Sponges And Placozoa
Sponges And  PlacozoaSponges And  Placozoa
Sponges And PlacozoaRitika Hamal
 
Presentacion visita al museo sorolla
Presentacion visita al museo sorollaPresentacion visita al museo sorolla
Presentacion visita al museo sorollarociogarpor72
 
I BUS 490 - P&G Team Packaging Case
I BUS 490 - P&G Team Packaging CaseI BUS 490 - P&G Team Packaging Case
I BUS 490 - P&G Team Packaging Casedevonridley
 
Survey Of Protozoans
Survey Of ProtozoansSurvey Of Protozoans
Survey Of ProtozoansRitika Hamal
 
TrustMe - Concorso Telecom-Sinfonia
TrustMe - Concorso Telecom-SinfoniaTrustMe - Concorso Telecom-Sinfonia
TrustMe - Concorso Telecom-Sinfoniagiuseppe_ravida
 
2011 annual dinner meeting presentation
2011 annual dinner meeting presentation2011 annual dinner meeting presentation
2011 annual dinner meeting presentationHutchIT
 
Analisi di un protocollo per la negoziazione di SLA in ambienti SOA.
Analisi di un protocollo per la negoziazione di SLA in ambienti SOA.Analisi di un protocollo per la negoziazione di SLA in ambienti SOA.
Analisi di un protocollo per la negoziazione di SLA in ambienti SOA.giuseppe_ravida
 
Sponges And Placozoa(2)
Sponges And  Placozoa(2)Sponges And  Placozoa(2)
Sponges And Placozoa(2)Ritika Hamal
 
Frigoriste
FrigoristeFrigoriste
Frigoristeeihaddad
 
Abiotic Effects(2)
Abiotic Effects(2)Abiotic Effects(2)
Abiotic Effects(2)Ritika Hamal
 
Population Characteristics
Population CharacteristicsPopulation Characteristics
Population CharacteristicsRitika Hamal
 

En vedette (20)

Derpless Discourse: authentic promotion for responding to Forrester, Gartner,...
Derpless Discourse: authentic promotion for responding to Forrester, Gartner,...Derpless Discourse: authentic promotion for responding to Forrester, Gartner,...
Derpless Discourse: authentic promotion for responding to Forrester, Gartner,...
 
profile.
profile.profile.
profile.
 
Abiotic Effects
Abiotic EffectsAbiotic Effects
Abiotic Effects
 
Premiers pas
Premiers pasPremiers pas
Premiers pas
 
Integers
IntegersIntegers
Integers
 
Sponges And Placozoa
Sponges And  PlacozoaSponges And  Placozoa
Sponges And Placozoa
 
Presentacion visita al museo sorolla
Presentacion visita al museo sorollaPresentacion visita al museo sorolla
Presentacion visita al museo sorolla
 
Millenials at work
Millenials at workMillenials at work
Millenials at work
 
o-net 54
o-net 54o-net 54
o-net 54
 
I BUS 490 - P&G Team Packaging Case
I BUS 490 - P&G Team Packaging CaseI BUS 490 - P&G Team Packaging Case
I BUS 490 - P&G Team Packaging Case
 
Survey Of Protozoans
Survey Of ProtozoansSurvey Of Protozoans
Survey Of Protozoans
 
TrustMe - Concorso Telecom-Sinfonia
TrustMe - Concorso Telecom-SinfoniaTrustMe - Concorso Telecom-Sinfonia
TrustMe - Concorso Telecom-Sinfonia
 
2011 annual dinner meeting presentation
2011 annual dinner meeting presentation2011 annual dinner meeting presentation
2011 annual dinner meeting presentation
 
Bug XTracker Report
Bug XTracker ReportBug XTracker Report
Bug XTracker Report
 
Analisi di un protocollo per la negoziazione di SLA in ambienti SOA.
Analisi di un protocollo per la negoziazione di SLA in ambienti SOA.Analisi di un protocollo per la negoziazione di SLA in ambienti SOA.
Analisi di un protocollo per la negoziazione di SLA in ambienti SOA.
 
Sponges And Placozoa(2)
Sponges And  Placozoa(2)Sponges And  Placozoa(2)
Sponges And Placozoa(2)
 
Frigoriste
FrigoristeFrigoriste
Frigoriste
 
Abiotic Effects(2)
Abiotic Effects(2)Abiotic Effects(2)
Abiotic Effects(2)
 
Biogeochemical
BiogeochemicalBiogeochemical
Biogeochemical
 
Population Characteristics
Population CharacteristicsPopulation Characteristics
Population Characteristics
 

Similaire à Fulltext

Large Scale Kernel Learning using Block Coordinate Descent
Large Scale Kernel Learning using Block Coordinate DescentLarge Scale Kernel Learning using Block Coordinate Descent
Large Scale Kernel Learning using Block Coordinate DescentShaleen Kumar Gupta
 
BARRACUDA, AN OPEN SOURCE FRAMEWORK FOR PARALLELIZING DIVIDE AND CONQUER ALGO...
BARRACUDA, AN OPEN SOURCE FRAMEWORK FOR PARALLELIZING DIVIDE AND CONQUER ALGO...BARRACUDA, AN OPEN SOURCE FRAMEWORK FOR PARALLELIZING DIVIDE AND CONQUER ALGO...
BARRACUDA, AN OPEN SOURCE FRAMEWORK FOR PARALLELIZING DIVIDE AND CONQUER ALGO...IJCI JOURNAL
 
IRJET- Latin Square Computation of Order-3 using Open CL
IRJET- Latin Square Computation of Order-3 using Open CLIRJET- Latin Square Computation of Order-3 using Open CL
IRJET- Latin Square Computation of Order-3 using Open CLIRJET Journal
 
VauhkonenVohraMadaan-ProjectDeepLearningBenchMarks
VauhkonenVohraMadaan-ProjectDeepLearningBenchMarksVauhkonenVohraMadaan-ProjectDeepLearningBenchMarks
VauhkonenVohraMadaan-ProjectDeepLearningBenchMarksMumtaz Hannah Vauhkonen
 
On Implementation of Neuron Network(Back-propagation)
On Implementation of Neuron Network(Back-propagation)On Implementation of Neuron Network(Back-propagation)
On Implementation of Neuron Network(Back-propagation)Yu Liu
 
interfacing matlab with embedded systems
interfacing matlab with embedded systemsinterfacing matlab with embedded systems
interfacing matlab with embedded systemsRaghav Shetty
 
Performance comparison of row per slave and rows set per slave method in pvm ...
Performance comparison of row per slave and rows set per slave method in pvm ...Performance comparison of row per slave and rows set per slave method in pvm ...
Performance comparison of row per slave and rows set per slave method in pvm ...eSAT Journals
 
Performance comparison of row per slave and rows set
Performance comparison of row per slave and rows setPerformance comparison of row per slave and rows set
Performance comparison of row per slave and rows seteSAT Publishing House
 
Parallel Machine Learning
Parallel Machine LearningParallel Machine Learning
Parallel Machine LearningJanani C
 
FAST AND EFFICIENT IMAGE COMPRESSION BASED ON PARALLEL COMPUTING USING MATLAB
FAST AND EFFICIENT IMAGE COMPRESSION BASED ON PARALLEL COMPUTING USING MATLABFAST AND EFFICIENT IMAGE COMPRESSION BASED ON PARALLEL COMPUTING USING MATLAB
FAST AND EFFICIENT IMAGE COMPRESSION BASED ON PARALLEL COMPUTING USING MATLABJournal For Research
 
An Optimized Parallel Algorithm for Longest Common Subsequence Using Openmp –...
An Optimized Parallel Algorithm for Longest Common Subsequence Using Openmp –...An Optimized Parallel Algorithm for Longest Common Subsequence Using Openmp –...
An Optimized Parallel Algorithm for Longest Common Subsequence Using Openmp –...IRJET Journal
 

Similaire à Fulltext (20)

A0270107
A0270107A0270107
A0270107
 
Large Scale Kernel Learning using Block Coordinate Descent
Large Scale Kernel Learning using Block Coordinate DescentLarge Scale Kernel Learning using Block Coordinate Descent
Large Scale Kernel Learning using Block Coordinate Descent
 
BARRACUDA, AN OPEN SOURCE FRAMEWORK FOR PARALLELIZING DIVIDE AND CONQUER ALGO...
BARRACUDA, AN OPEN SOURCE FRAMEWORK FOR PARALLELIZING DIVIDE AND CONQUER ALGO...BARRACUDA, AN OPEN SOURCE FRAMEWORK FOR PARALLELIZING DIVIDE AND CONQUER ALGO...
BARRACUDA, AN OPEN SOURCE FRAMEWORK FOR PARALLELIZING DIVIDE AND CONQUER ALGO...
 
IRJET- Latin Square Computation of Order-3 using Open CL
IRJET- Latin Square Computation of Order-3 using Open CLIRJET- Latin Square Computation of Order-3 using Open CL
IRJET- Latin Square Computation of Order-3 using Open CL
 
VauhkonenVohraMadaan-ProjectDeepLearningBenchMarks
VauhkonenVohraMadaan-ProjectDeepLearningBenchMarksVauhkonenVohraMadaan-ProjectDeepLearningBenchMarks
VauhkonenVohraMadaan-ProjectDeepLearningBenchMarks
 
FrackingPaper
FrackingPaperFrackingPaper
FrackingPaper
 
On Implementation of Neuron Network(Back-propagation)
On Implementation of Neuron Network(Back-propagation)On Implementation of Neuron Network(Back-propagation)
On Implementation of Neuron Network(Back-propagation)
 
Group Project
Group ProjectGroup Project
Group Project
 
interfacing matlab with embedded systems
interfacing matlab with embedded systemsinterfacing matlab with embedded systems
interfacing matlab with embedded systems
 
Performance comparison of row per slave and rows set per slave method in pvm ...
Performance comparison of row per slave and rows set per slave method in pvm ...Performance comparison of row per slave and rows set per slave method in pvm ...
Performance comparison of row per slave and rows set per slave method in pvm ...
 
Performance comparison of row per slave and rows set
Performance comparison of row per slave and rows setPerformance comparison of row per slave and rows set
Performance comparison of row per slave and rows set
 
Jf3515881595
Jf3515881595Jf3515881595
Jf3515881595
 
Computation Assignment Help
Computation Assignment Help Computation Assignment Help
Computation Assignment Help
 
Matlab ppt
Matlab pptMatlab ppt
Matlab ppt
 
Parallel Machine Learning
Parallel Machine LearningParallel Machine Learning
Parallel Machine Learning
 
FAST AND EFFICIENT IMAGE COMPRESSION BASED ON PARALLEL COMPUTING USING MATLAB
FAST AND EFFICIENT IMAGE COMPRESSION BASED ON PARALLEL COMPUTING USING MATLABFAST AND EFFICIENT IMAGE COMPRESSION BASED ON PARALLEL COMPUTING USING MATLAB
FAST AND EFFICIENT IMAGE COMPRESSION BASED ON PARALLEL COMPUTING USING MATLAB
 
Calfem34
Calfem34Calfem34
Calfem34
 
An Optimized Parallel Algorithm for Longest Common Subsequence Using Openmp –...
An Optimized Parallel Algorithm for Longest Common Subsequence Using Openmp –...An Optimized Parallel Algorithm for Longest Common Subsequence Using Openmp –...
An Optimized Parallel Algorithm for Longest Common Subsequence Using Openmp –...
 
An Introduction to MATLAB with Worked Examples
An Introduction to MATLAB with Worked ExamplesAn Introduction to MATLAB with Worked Examples
An Introduction to MATLAB with Worked Examples
 
Gk3611601162
Gk3611601162Gk3611601162
Gk3611601162
 

Dernier

Measures of Central Tendency: Mean, Median and Mode
Measures of Central Tendency: Mean, Median and ModeMeasures of Central Tendency: Mean, Median and Mode
Measures of Central Tendency: Mean, Median and ModeThiyagu K
 
Ecosystem Interactions Class Discussion Presentation in Blue Green Lined Styl...
Ecosystem Interactions Class Discussion Presentation in Blue Green Lined Styl...Ecosystem Interactions Class Discussion Presentation in Blue Green Lined Styl...
Ecosystem Interactions Class Discussion Presentation in Blue Green Lined Styl...fonyou31
 
Web & Social Media Analytics Previous Year Question Paper.pdf
Web & Social Media Analytics Previous Year Question Paper.pdfWeb & Social Media Analytics Previous Year Question Paper.pdf
Web & Social Media Analytics Previous Year Question Paper.pdfJayanti Pande
 
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptx
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptxPOINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptx
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptxSayali Powar
 
Student login on Anyboli platform.helpin
Student login on Anyboli platform.helpinStudent login on Anyboli platform.helpin
Student login on Anyboli platform.helpinRaunakKeshri1
 
Arihant handbook biology for class 11 .pdf
Arihant handbook biology for class 11 .pdfArihant handbook biology for class 11 .pdf
Arihant handbook biology for class 11 .pdfchloefrazer622
 
Grant Readiness 101 TechSoup and Remy Consulting
Grant Readiness 101 TechSoup and Remy ConsultingGrant Readiness 101 TechSoup and Remy Consulting
Grant Readiness 101 TechSoup and Remy ConsultingTechSoup
 
Introduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The BasicsIntroduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The BasicsTechSoup
 
Z Score,T Score, Percential Rank and Box Plot Graph
Z Score,T Score, Percential Rank and Box Plot GraphZ Score,T Score, Percential Rank and Box Plot Graph
Z Score,T Score, Percential Rank and Box Plot GraphThiyagu K
 
Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...
Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...
Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...Krashi Coaching
 
JAPAN: ORGANISATION OF PMDA, PHARMACEUTICAL LAWS & REGULATIONS, TYPES OF REGI...
JAPAN: ORGANISATION OF PMDA, PHARMACEUTICAL LAWS & REGULATIONS, TYPES OF REGI...JAPAN: ORGANISATION OF PMDA, PHARMACEUTICAL LAWS & REGULATIONS, TYPES OF REGI...
JAPAN: ORGANISATION OF PMDA, PHARMACEUTICAL LAWS & REGULATIONS, TYPES OF REGI...anjaliyadav012327
 
1029 - Danh muc Sach Giao Khoa 10 . pdf
1029 -  Danh muc Sach Giao Khoa 10 . pdf1029 -  Danh muc Sach Giao Khoa 10 . pdf
1029 - Danh muc Sach Giao Khoa 10 . pdfQucHHunhnh
 
CARE OF CHILD IN INCUBATOR..........pptx
CARE OF CHILD IN INCUBATOR..........pptxCARE OF CHILD IN INCUBATOR..........pptx
CARE OF CHILD IN INCUBATOR..........pptxGaneshChakor2
 
BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...
BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...
BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...Sapna Thakur
 
mini mental status format.docx
mini    mental       status     format.docxmini    mental       status     format.docx
mini mental status format.docxPoojaSen20
 
Beyond the EU: DORA and NIS 2 Directive's Global Impact
Beyond the EU: DORA and NIS 2 Directive's Global ImpactBeyond the EU: DORA and NIS 2 Directive's Global Impact
Beyond the EU: DORA and NIS 2 Directive's Global ImpactPECB
 
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptx
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptxSOCIAL AND HISTORICAL CONTEXT - LFTVD.pptx
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptxiammrhaywood
 
APM Welcome, APM North West Network Conference, Synergies Across Sectors
APM Welcome, APM North West Network Conference, Synergies Across SectorsAPM Welcome, APM North West Network Conference, Synergies Across Sectors
APM Welcome, APM North West Network Conference, Synergies Across SectorsAssociation for Project Management
 
Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...
Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...
Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...EduSkills OECD
 

Dernier (20)

Measures of Central Tendency: Mean, Median and Mode
Measures of Central Tendency: Mean, Median and ModeMeasures of Central Tendency: Mean, Median and Mode
Measures of Central Tendency: Mean, Median and Mode
 
Ecosystem Interactions Class Discussion Presentation in Blue Green Lined Styl...
Ecosystem Interactions Class Discussion Presentation in Blue Green Lined Styl...Ecosystem Interactions Class Discussion Presentation in Blue Green Lined Styl...
Ecosystem Interactions Class Discussion Presentation in Blue Green Lined Styl...
 
Web & Social Media Analytics Previous Year Question Paper.pdf
Web & Social Media Analytics Previous Year Question Paper.pdfWeb & Social Media Analytics Previous Year Question Paper.pdf
Web & Social Media Analytics Previous Year Question Paper.pdf
 
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptx
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptxPOINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptx
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptx
 
Student login on Anyboli platform.helpin
Student login on Anyboli platform.helpinStudent login on Anyboli platform.helpin
Student login on Anyboli platform.helpin
 
Arihant handbook biology for class 11 .pdf
Arihant handbook biology for class 11 .pdfArihant handbook biology for class 11 .pdf
Arihant handbook biology for class 11 .pdf
 
Grant Readiness 101 TechSoup and Remy Consulting
Grant Readiness 101 TechSoup and Remy ConsultingGrant Readiness 101 TechSoup and Remy Consulting
Grant Readiness 101 TechSoup and Remy Consulting
 
Introduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The BasicsIntroduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The Basics
 
Z Score,T Score, Percential Rank and Box Plot Graph
Z Score,T Score, Percential Rank and Box Plot GraphZ Score,T Score, Percential Rank and Box Plot Graph
Z Score,T Score, Percential Rank and Box Plot Graph
 
Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...
Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...
Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...
 
JAPAN: ORGANISATION OF PMDA, PHARMACEUTICAL LAWS & REGULATIONS, TYPES OF REGI...
JAPAN: ORGANISATION OF PMDA, PHARMACEUTICAL LAWS & REGULATIONS, TYPES OF REGI...JAPAN: ORGANISATION OF PMDA, PHARMACEUTICAL LAWS & REGULATIONS, TYPES OF REGI...
JAPAN: ORGANISATION OF PMDA, PHARMACEUTICAL LAWS & REGULATIONS, TYPES OF REGI...
 
1029 - Danh muc Sach Giao Khoa 10 . pdf
1029 -  Danh muc Sach Giao Khoa 10 . pdf1029 -  Danh muc Sach Giao Khoa 10 . pdf
1029 - Danh muc Sach Giao Khoa 10 . pdf
 
CARE OF CHILD IN INCUBATOR..........pptx
CARE OF CHILD IN INCUBATOR..........pptxCARE OF CHILD IN INCUBATOR..........pptx
CARE OF CHILD IN INCUBATOR..........pptx
 
Código Creativo y Arte de Software | Unidad 1
Código Creativo y Arte de Software | Unidad 1Código Creativo y Arte de Software | Unidad 1
Código Creativo y Arte de Software | Unidad 1
 
BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...
BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...
BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...
 
mini mental status format.docx
mini    mental       status     format.docxmini    mental       status     format.docx
mini mental status format.docx
 
Beyond the EU: DORA and NIS 2 Directive's Global Impact
Beyond the EU: DORA and NIS 2 Directive's Global ImpactBeyond the EU: DORA and NIS 2 Directive's Global Impact
Beyond the EU: DORA and NIS 2 Directive's Global Impact
 
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptx
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptxSOCIAL AND HISTORICAL CONTEXT - LFTVD.pptx
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptx
 
APM Welcome, APM North West Network Conference, Synergies Across Sectors
APM Welcome, APM North West Network Conference, Synergies Across SectorsAPM Welcome, APM North West Network Conference, Synergies Across Sectors
APM Welcome, APM North West Network Conference, Synergies Across Sectors
 
Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...
Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...
Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...
 

Fulltext

  • 1. Purdue University Purdue e-Pubs ECE Technical Reports Electrical and Computer Engineering 9-1-1992 Implementation of back-propagation neural networks with MatLab Jamshid Nazari Purdue University School of Electrical Engineering Okan K. Ersoy Purdue University School of Electrical Engineering This document has been made available through Purdue e-Pubs, a service of the Purdue University Libraries. Please contact epubs@purdue.edu for additional information. Nazari, Jamshid and Ersoy, Okan K., "Implementation of back-propagation neural networks with MatLab" (1992). ECE Technical Reports. Paper 275. http://docs.lib.purdue.edu/ecetr/275
  • 3. TABLE OF CONTENTS Page LIST OF TABLES . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . v LIST OF FIGURES . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . vi ABSTRACT . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . vii 1. BACK PROPAGATION ALGORITHM USING MATLAB . . . . . . . . 1 What is Matlab? . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1 Why Use Matlab? . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2 Speed Comparison of Matrix Multiply in Matlab and C . . . . . . . 2 Back Propagation Algorithm . . . . . . . . . . . . . . . . . . . . . . . 3 Mbackprop Program . . . . . . . . . . . . . . . . . . . . . . . . . . . 3 Reducing Number of Iterations Increases Execution Speed . . . . . 4 Speed Comparison of Algorithm 1and Algorithm 2 . . . . . . . . . 5 Matlab Backprop Speed vs. C Backprop Speed . . . . . . . . . . . . 7 Integrated Graphical Capability of the Mbackprop Program . . . . . 9 Other Capabilities of the Mbackprop Package . . . . . . . . . . . . . . 10 Summary and Conclusions . . . . . . . . . . . . . . . . . . . . . . . . 11 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .BIBLIOGRAPHY I!,
  • 4. LIST OF TABLES Page Speed Comparison of Matrix Multiply in Matlab and a C program . . 2 Algorithm 1 Solves Class Identification Problem . . . . . . . . . . . . 6 . . . . . . . . . . . .Algorithm 2 Solves Class Identification Problem 7 . . . . . . . . . . . . . . .Size of the Variables in Algorithms1 and 2 8 Speed of Algorithm1 vs. Speed of Algorithm 2 . . . . . . . . . . . . 9 Speed of Matlab Backprop Program vs. Speed of C Backprop Programs 9 . . .Comparison of Speeds in Single and Double Precision Backprops 10
  • 5. LIST OF FIGURES Figure 1.1 1.2 1.3 1.4 1.5 1.6 1.7 I't~gt- . . . . . . . . . . . . . . . . . . . . . .Variables Used in Algorithm 1 12 . . . . . . . . . . . . . . . . . . . . . .Variables Used in Algorithm 2 13 Sample Mean Square Error Graph Generated by Mbackprop . . . . . 14 . . . . . . .Sample Percent Correct Graph Generated by Mbackprop 15 Sample Maximum Absolute Error Graph Generated by Mbackprop . . 16 . . . . .Sample Percent Bits Wrong Graph Generated by Mbackprop 17 . . . . . . . . . . .Sample Compact Graph Generated by Mbackprop 18
  • 6. ABSTRACT The artificial neural network back propagation algorithm is implemented in Mat- lab language. This implementation is compared with severalother software packages. The effect of reducing the number of iterations in the performance of the algorithm iai studied. The speed of the back propagation program, mkckpmp, written in Mat- lab language is compared with the speed of several other back propagation programs which are written in the C language. The speed of the Matlab program mbackpmp is,also compared with the C program quickpmp which is a variant of the back prop- agation algorithm. It is shown that the Matlab program mbackpmp is about 4.5 to 7 times faster than the C programs.
  • 7. 1. BACK PROPAGATION ALGORITHM USING MATLAB This chapter explains the software package, mbackprop, which is written in MatJah language. The package implements the Back Propagation (BP) algorithm [RIIW861, which is an artificial neural network algorithm. There are other software packages which implement the back propagation algo- rithm. For example the AspirinIMIGRAINESSoftware Tools [Leig'I]is intended to be used to investigate different neural network paradigms. There is also NASA NETS [Baf89] which is a neural network simulator. It provides a system for a variety of neural network configurations which uses generalized delta back propagation learn- ing method. There are also books which have implementation of BP algorithm in C language for example, see [ED90]. Many of these software packages are huge, they need to be compiled and some- times difficult to understand. Modification of these codes requires understanding the rnassive amount of source code and additional low level programming. The mbackprop on the other hand is easy to use and very fast. With the graphical capability of the Idatlab the network parameters can be graphed to see what is going on inside any specific network. Additions and modifications to the mbackprop package are easier a~ndfurther research in the area of neural network can be facilitated. 1.1 What is Matlab? Matlab is a commercial software developed by Mathworks Inc. It is an interactive software package for scientific and engineering numeric computation [Inc90]. Matlab has several basic routines which do matrix arithmetics, plotting etc.
  • 8. 1..2 Why Use Matlab? Matlab is already in use in many institutions. It is used in research in academia and industry. Prototype solutions are usually obtained faster in Matlab than solving a, problem from a programming language. Matlab is fast, because the core routines in Matlab are fine tuned for diflerent computer architectures. Following test was made to compare the speed between Matlab and a program written in C. Since the back propagation algorithm involves nnatrix manipulations the test chosen was matrix multiply. As the next section shows, ndatlab1 was about 2.5 times faster than a C program both doing a matrix multiply. !?:peedComparison of Matrix Multiply in Matlab and C A program in C was written to multiply two matrices containing double precision numbers. The result of the multiplication is assigned into a third matrix. Each matrix contained 500 rows and 500 columns. A Matlab M file was written to do the same multiply as C program did. Only the segment of the code which does the nlultiplication is timed. The test was run on an IPC-SparcStation computer, the rlesult is shown in Table 1.1. As the table shows Matlab is faster than the C program bly more than a factor of two. Table 1.1 Speed comparison of matrix multiply in Matlab and a C program. Matlab runs 2.5 times faster than the C program. 'The version of Matlab we used waa 3.5i.
  • 9. 1..3 Back Propagation Algorithm The generalized delta rule [RHWSG], also known as back propagation algorit,li~n is explained here briefly for feed forward Neural Network (NN). The explanitt,ion Ilcrc is intended to give an outline of the process involved in back propagation algorithm. The NN explained here contains three layers. These are input, hidden, and output Layers. During the training phase, the training data is fed into to the input layer. The dlata is propagated to the hidden layer and then to the output layer. This is called the forward pass of the back propagation algorithm. In forward pass, each node in hidden layer gets input from all the nodes from input layer, which are multiplied with appropriate weights and then summed. The output of the hidden node is the non- linear transformation of the this resulting sum. Similarly each node in output layer gets input from all the nodes from hidden layer, which are multiplied with appropriate weights and then summed. The output of this node is the non-linear transformation of the resulting sum. The output values of the output layer are compared with the target output values. The target output values are those that we attempt to teach our network. The error between actual output values and target output values is calculated and propagated back toward hidden layer. This is called the backward pass of the back propagation algorithm. The error is used to update the connection strengths between nodes, i.e. weight matrices between input-hidden layers and hidden-output layers are updated. During the testing phase, no learning takes place i.e., weight matrices are not changed. Each test vector is fed into the input layer. The feed forward of the testing data is similar to the feed forward of the training data. 1..4 Mbackprop Program The mbackprop program is written in Matlab language. The program implements the back propagation algorithm [RHW86]. The algorithms used in the nabackprop program involve very few number of iterations. This is one of the reasons why this
  • 10. program is so fast. In the next section, an example is given to see the effect of reducing number of iterations has on the executionspeed of a program. In Section1.5execution speed of the mbackprop program in Matlab is compared with the execution speed of a back propagation program in C. fiducing Number of Iterations Increases Execution Spccd There are several ways to write a program to accomplish a given task. The approach or algorithm a person might take will have a great effect on the execution s p d of a program. Here, a class identification problem is stated and then two solutions are presented. Statement of the problem is, given a matrix A, find the class to which each column of the matrix A belongs. Each column of the matrix A is a vector x which we want tofind to which class this veztor belongs. To do this, for each of these vectors x, we want to find the distances between the vector x and m other vectors. These m vectors are the desired vectors representing classl through class,. The minimum of the m distances, comes from a vector representing class,. The number j is the answer to the column vector x. So the desired output is a row vector B indicating to which class each of the vectors in A belongs. Content of the matrix A is changing, so we need to calculate the row vector B more than once. Two solutions are now presented for the above problem. The first solution will be algorithm 1 and the second solution will be algorithm 2. Both algorithms will need az~input argument the following variables: variable "A" which contains the matrix A variable "Classes" which contains vectors representing classr through class, variable "nClassesn which contains the number of classes rn. The output of the both algorithms is variable "B" which will contain the class number of each column of the variable "A". Figure1.1shows severalof the variablesused in algorithm 1. Here the variable "An
  • 11. is made of columns XI, 2 2 , ...,x,. Variable 'Classes" is made of columns cl,c2,...,cm which represents classl through class,. Variable 'dist" is a column vector of size m which will hold the distance of a vector x in A to each of the m classes in variable "Classesn . The algorithm 1 is the following: for each xi in A where i = 1,...,n - dist(j) = Square Euclidean Distance( x;, c; ) where j = 1,...,m - B(i) = k where dist(k) = min( dist ) Figure 1.2 shows several of the variables used in algorithm 2. Here the variable 'An is also made of columns xl,x2,...,x, but we will view it as one block. Variable 'Classesn is made of m column blocks e l , ...,C,, where m is the number of classes. Each block Cj is the same size as block A. The block C; contains n equal columns where n is the number of columns in A. Each column in block Cj is cj which represents classj. Variable "distn is made of m block columns which will hold the distance of bllock A to each of the m blocks in variable uClasses". The algorithm 2 is the following: dist(j,:) = Square Euclidean Distance( A, Cj ) where j = l!,...,m. dist(j,:) refers to row j of dist matrix. B(i) = k where dist(k,i) = min( did(:, i) ). did(:, i) refers to column i of dist matrix. Tables 1.2 and 1.3 show the two solutions for the class identification problem using algorithms 1 and 2. Note that these solutions are written in 'Matlab language. The algorithm 1 used in Table 1.2 is straight forward. As shown in the next section, the algorithm 1 contains much more iterations than algorithm 2. This causing the aJgorithm 1 to run slower than the algorithm 2 of Table 1.3. Speed Comparison of Algorithm 1 and Algorithm 2 The above algorithms were used to solve the class identification problem, where the number of classes was 8. The size of the variables used in algorithms 1 and 2
  • 12. Table 1.2 Algorithm 1 is a straight forward method which solves the class identifica- tion problem. 1) function B=algorithml( A, Clseeerr, nClasses ) 2) % Each column of the variable "Clesseenrepresents a class 3) [ n b w , nCol ] = size( A ); 4) B = sem(1, nCol ); % Preallocate memory 5) for i = 1: nCol, 6) x = A ( : , i ) ; 7) for j = 1 : nclaeses, 8) dist( j ) = sum( ( x - Claeses( :,j ) ) .- 2 ); 9) end 10) [ v, B(i) ] = min( dist ); 11) end are shown in Table 1.4. Note that the amount of memory used by algorithm 2 (1922 Kbytes) is muchgreater than the memory used in algorithm 1 (212 Kbytes). However, as shown below, algorithm 2 is much faster than algorithm 1. The speed of execution is related to the number of iterations in the algorithm. The number of iterations for algorithm 1 is much greater than the number of iterations for algorithm 2. In this example, the statement number 8 in algorithm 1 gets executed 24,000 (nCol x nClasses = 3000 x 8) times. Where in algorithm 2 either statement number 11 or 15 gets executed only 8 (nclasses = 8) times. Since hdatlab is an interpretive language algorithm 1 is much slower than algorithm 2. Table 1.5 shows that algorithm 2 runs about 23 times faster than algorithm 1. The test waa performed on an IPC-SparcStation computer. In the next section the speed of mbackpmp program, written in Matlab, is compared to the speed of a C program bloth implementing the back propagation algorithm [RHW861.
  • 13. Table 1.3 Algorithm 2 is another way to solve the class identification problem. It is :faster than Algorithm 1. 1) function B=algorithm2( A, Classes, nClasses ) 2) % Each column of the variable "Classesn represents all of the 3) % "nclasses" classes. If there are 8 clasaea and each class is 4) % represented by 8 numbers, then the number of rows of "Classes" is 5) % equal to 64. The number of columns in "Classes" is equal to number 6) % of columns in A. 7) [ n b w , nCol ] = size( A ); 8) dist = zeros( nclasses, nCol ); % Preallocate memory 9) if n b w == 1 10) for j = 1 : nclasses, 11) dist( j, : ) = ( A - Classes( j, : ) ) .' 2; 12) end 13) else 14) forj=l:nClasses, 15) dist(j,:)=sum((A-Classes(((j-l)*nRow+l):(j*nRow),:)) . - 2 ) ; 16) end 17) end 18) [ v, B ] = min( dist ); yll.5 Matlab Backprop Speed vs. C Backprop Speed The back propagation program in Matlab, mbackprop, is compared with two other (Z back propagation programs fbackprop2 and dbackprop. The mbackprop is also com- pared with the C program quickprop [Fah88]. The quickprop program is a modification of a back propagation program which has similar feed forward and backward routines but in update weight routine, all the weights are updated as a function of each weight's current slope, previous slope, and the size of the last jump. However, if the variable "ModeSwitchThreshold" in the quickprop program is set to a big number then all the weight updates are based on normal gradient descent method i.e. same as in regular back propagation algorithm. The program fbackprop is similar to the program dbackprop. The only differenceis that the calculations in fbackpmpare in floats (single precision), where the calculations aThe jbaekprop program has been used in some of our research at Purdue University for last two years.
  • 14. 'I'able 1.4 Size of the variables in algorithms 1 and 2 is shown here. The amount of memory used in algorithm 1 is 212 Kbytes where algorithm 2 uses 1922 Kbytcs. in dbackpropare in doubles (double precision). All the calculations in Matlab program Variable Name A B Clasees dist I j nClasea nCol n h w v x tirzbackpropare in doubles. The calculations in the quickprop program are in floats. In the fbackprop and the dbackprop programs, weights get updated after every in- Siee of Variable in Algorithm 1 8 x 3000 1 x 3000 8 x 8 8 x 1 1 x 1 1 x 1 1 x 1 1 x 1 1 x 1 1 x 1 8 x 1 p,ut/output vector pair. Where the weights in quickpmp and mbackprop programs get updated after a complete sweep of the training data. As shown below the mbackprop Siw of Variable in Algorithm 2 8 x 3000 1 x 3000 64 x 3000 8 x 3000 1 x 1 1 x 1 1 x 1 1 x 1 1 x 3000 program is faster than all the three C programs. The neural network, used in our benchmark tests, had 64 input nodes, 16 hidden #Bytea Alg 1 192,000 24,000 512 64 8 8 8 8 8 8 64 nodes, and 8 output nodes. The training data contained 1600 input and output vector pairs. Each input vector was 64 numbers and each output vector was 8 numbers. The #Bytes Alg 2 192,000 24,000 1,536,000 192,000 8 8 8 8 24,000 training timefor 100 sweepsover the training data is measured for the above programs using an IPC-SparcStation Computer. The results are shown in Table 1.6. As the tableshows, the mbackprop runs 7.0 times faster than the C program dbackprop3 . The - mbackprop runs 4.5 times faster than the C program quickprop. The training time for the C programs j5ackprop4 and dbackprop is also measured in two other computerss. One of the computers was a Vax 11/780 and the other 3The training time for jhockpmp, dbockpmp, and quickpmp are measured for 10 iterations. To get the training time for 100 iterations, the measured nurnbcrcl are multiplied by 10. h he training time for fbackpmp and dbackpmp are measured for 1 iteration. To get the training time for 100 iterations, the measured numbers are multiplied by 100. 6Wedid not have Matlab running in thee computers
  • 15. Table 1.5 Speed of algorithm 1 is compared to the speed of algorithm 2. Algorithm 2 runs about 23 times faster than algorithm 1. Table 1.6 Speed of the Matlab program mbackprop is compared to the speed of C programs fbackprop, dbackprop and quickprop. The training time for 100 swcvps ovcr the training data is measured. The Matlab program mbackprop runs 7.0 times faster than the C program dbackprop. The mbackprop also runs 4.5 times faster than the quickprop program. Program Execution Time fbackprop 3969 seconds dbackprop 3792 seconds mbackprop 536 seconds quickprop 2407 secondsi computer was a Zenith 386133 running SCO unix with math coprocessor. Table 1.7 ishows the training time for 100 iterations over the training data. The time taken for the fbackpropprogram was less than the dbackprop program in these computers, how- ever the fbackprop time of the IPC-SparcStation computer was longer the dbackprop time. So depending on different computer architectures floating point single precision calculations are faster than double precision calculations or vice versa. As it was shown, the mbackprop program was fastest among the programs we considered above. However mbackprop provides an integrated graphic capability that other programs lack. 1..6 Integrated Graphical Capability of the Mbackprop Program The back propagation program mbackprop is faster than the C programs consid- ered here. However this is not the only advantage that this program has over others. It provides an integrated graphical capability and an interprative environment that other programs lack.
  • 16. Table 1.7 Execution speed of the fbackprop, a single precision back propagation pro- gram, is compared to the dbackprop a double precision back propagation prograrrr. The training time for 100 sweeps over the training data is nlca..urc:d o ~ lI,hru: co~rl- puters. In the IPC-SparcStation computer double precision program was fmtc?rthan t'he single precision program. In Vax 11/780 and 386 computer the single precision program is faster than the double precision program. The execrltion time of the rnbackprop is included here for comparison. In the mbackprop program, the network parameters can be easily viewed during program execution. Training and testing reports can be enabled during training and statistics such as mean square error, percent correct, etc. can be collected in report intervals specified by the user. Figure 1.3 shows a sample 'mean square error'graph. Figure 1.4 shows a sample 'percent correct'graph. Percent correct refers to the percentage of the input vectors, in training or testing data, which are correctly classified. Figure 1.5 shows a sample 'mazimum absolute e m r ' graph. Figure 1.6 shows a sample 'percent bits wrong' graph. Percent bits wrong refers to the percentage of the output nodes which were off more than some threshold. Figure 1.7 shows a sample 'compact graph' graph. Compact graph is a graph which contains the above 4 graphs in one graph. Other graphs can be easily added to the mbackprop package. In the next section vie list several other capabilities of the mbackprop program. Computer IPGSparcStation Vax 11/780 Zenith 386/33 1.7 Other Capabilities of the Mbackprop Package So far we haveshown that mbackproppackage is fast and contains several standard graphical capabilities. Several of the mbackprop capabilities are: Execution Time for fbaekprop 3,969 seconds 69,923 seconds 21,795seconds a allows a user to specify a weight file for initial weights to start training. Execution Time for Dbackprop 3,792 seconds 73,150 seconds 24,523seconds Execution Time for Mbackprop 536 seconds
  • 17. can generate random initial weights for training and allows the user to save these initial weights to be used later. if the training gets started in wrong initial weights, the program is easily inter- rupted and different set of initial weights is used. the result from training a network can be saved and recalled at a later time. allows further training from where it was last left off. l.8 Summary and Conclusions The mbackprop program is written in Matlab language. This program implements loheback propagation algorithm [RHW86]. Since Matlab is an interpretive language the number of iterations in the algorithms have been reduced. This reduced number of iterations results in a faster executable program. The mbackprop program is faster than the C back propagation program dbackprop by a factor of 7.0. It is faster than quickprop [Fah88] program by a factor of 4.5. The mbackprop provides other capabilities such as integrated graphics and interpretive environment which Matlab offers. The mbackprop size is less than a comparative program in C. It is modular and each individual module can be viewed as a software Integrated Chip (IC). Each software IC can be modified as long as the input/output criteria is met. Additions of other software ICs is easy to be incorporated into the mbackprop package. Further research in the area of neural network can be facilitated.
  • 18. Figure 1.1 Some of the variables used in algorithm 1 is shown here.
  • 19. Classes= dist = Figure 1.2 Some of the variables used in algorithm 2 is shown here.
  • 20. Figure1.3 A sample 'mean square emr'graph, generated by the mbackpropprogram, it3 shown here. The solid line is the training mean square error and the dashed line is the testing mean square error.
  • 21. Figure 1.4 A sample 'percent correct'graph, generated by the mbnckprop program, is shown here. The solid line is training percent correct and the dashed line is the testing percent correct.
  • 22. Figure 1.5 A sample 'maximum absolute error'graph, generated by the mbackprop program, is shown here. The solid line is training maximum absolute error and the dashed line is the testing maximumabsolute error.
  • 23. Figure 1.6 A sample 'percent bits wrong'graph, generated by the mbnckpropprogram, is shown here, The solid line is the training percent bits wrong and the dashed line is; the testing percent bits wrong.
  • 24. 0 ' I I 0 50 100 IterU 0' I I 0 50 100 IterU Figure 1.7 A sample 'compact 'graph, generated by the mbackprop program, is shown here. The solid lines refers to the training results and the dashed lines refer to the testing results.
  • 26. BIBLIOGRAPHY [Baf89] Paul T. Baffes. NETS Users's Guide, Version 2.0 of NETS. Technical Re- port JSC-23366, NASA, Software Technology Branch, Lyndon B. Johnson Space Center, September 1989. [ED901 R. C. Eberhart and R. W. Dobbins. Neuml Network PC Tools, A Practical Guide. Academic Press, San Diego, California 92101, 1990. [Fah88] E. Fahlman, Scott. An Empirical Study of Learning Speed in Back- Propagation Networks. Technical Report CMU-CS-88-162, CMU, CMU, September 1988. [Inc9OJ The MathWorks Inc. PRO-MATLAB for Sun Workstations, User's Guide. The MathWorks Inc., January 1990. [Lei911 Russell R. Leighton. The Aspirin/MIGRAINES Software Tools, User's Manual, Release V5.0. Technical Report MP-91W00050, MITRE Corpo- ration, MITRE Corporation, December 1991. [RHW86] D. E. Rumelhart, G. E. Hinton, and R. J. Williams. Leaning Internal Representaions by Ewvr Propagation in Rumelhart, D. E. and McClelland, J. L., Pamllel Distributed Processing: Explorations in the Microstructure of Cognition. MIT Press, Cambridge Massachusette, 1986.