SlideShare une entreprise Scribd logo
1  sur  18
Télécharger pour lire hors ligne
Computer Engineering and Intelligent Systems                                                  www.iiste.org
ISSN 2222-1719 (Paper) ISSN 2222-2863 (Online)
Vol 3, No.4, 2012

       Development of a Writer-Independent Online Handwritten
        Character Recognition System Using Modified Hybrid
                       Neural Network Model

                                                  Fenwa O. D.*
                               Department of Computer Science and Engineering,
                  Ladoke Akintola University of Technology, P.M.B 4000, Ogbomoso, Nigeria.
                         *E-mail of the corresponding author: odfenwa@lautech.edu.ng


                                                 Omidiora E. O.
                               Department of Computer Science and Engineering,
                  Ladoke Akintola University of Technology, P.M.B 4000, Ogbomoso, Nigeria.

                                       E-mail: omidiorasayo@yahoo.co.uk



                                                 Fakolujo O. A.
                               Department of Computer Science and Engineering,
                  Ladoke Akintola University of Technology, P.M.B 4000, Ogbomoso, Nigeria.

                                           E-mail: ola@fakolujo.com



                                                  Ganiyu R. A.
                               Department of Computer Science and Engineering,
                  Ladoke Akintola University of Technology, P.M.B 4000, Ogbomoso, Nigeria.

                                        E-mail: ganiyurafiu@yahoo.com


Abstract
Recognition of handwritten characters has become a difficult problem because of the high variability and
ambiguity in the character shapes written by individuals. Some of the problems encountered by researchers
include selection of efficient feature extraction method, long network training time, long recognition time
and low recognition accuracy. However, many feature extraction techniques have been proposed in
literature to improve overall recognition rate although most of the techniques used only one property of the
handwritten character. This research focuses on developing a feature extraction technique that combined
three characteristics (stroke information, contour pixels and zoning) of the handwritten character to create a
global feature vector. A hybrid feature extraction algorithm was developed to alleviate the problem of poor
feature extraction algorithm of online character recognition system. Besides, this research work also
focused on alleviating the problem of standard backpropagation algorithm based on ‘error adjustment’. A
hybrid of modified Counterpropagation and modified optical backpropagation neural network model was
developed to enhance the performance of the proposed character recognition system. Experiments were

                                                     89
Computer Engineering and Intelligent Systems                                                   www.iiste.org
ISSN 2222-1719 (Paper) ISSN 2222-2863 (Online)
Vol 3, No.4, 2012

performed with 6200 handwriting character samples (English uppercase, lowercase and digits) collected
from 50 subjects using G-Pen 450 digitizer and the system was tested with 100 character samples written
by people who did not participate in the initial data acquisition process. The performance of the system was
evaluated based on different learning rates, different image sizes and different database sizes. The
developed system achieved better performance with no recognition failure, 99% recognition rate and an
average recognition time of 2 milliseconds.


Keywords: Character recognition, Feature extraction, Neural Network, Counterpropagation, Optical
Backpropagation, Learning rate.

1. Introduction
The use of neural network for handwriting recognition is a field that is attracting a lot of attention. As the
computing technology advances, the benefits of using Artificial Neural Network (ANN) for the purpose of
handwriting recognition become more obvious. Hence, new ANN approaches geared toward the task of
handwriting recognition are constantly being studied. Character recognition is the process of applying
pattern-matching methods to character shapes that has been read into a computer to determine which alpha-
numeric character, punctuation marks, and symbols the shapes represent. The classes of recognition
systems that are usually distinguished are online systems for which handwriting data are captured during
the writing process (which makes available the information on the ordering of the strokes) and offline
systems for which recognition takes place on a static image captured once the writing process is over
(Anoop and Anil, 2004; Liu et al., 2004; Mohamad and Zafar, 2004; Naser et al., 2009; Pradeep et al.,
2011). The online methods have been shown to be superior to their offline counterpart in recognizing
handwriting characters due the temporal information available with the former (Pradeep et al., 2011).
Handwriting recognition system can further be broken down into two categories: writer independent
recognition system which recognizes wide range of possible writing styles and a writer dependent
recognition system which recognizes writing styles only from specific users (Santosh and Nattee, 2009).
Online handwriting recognition today has special interest due to increased usage of the hand held devices.
The incorporation of keyboard being difficult in the hand held devices demands for alternatives, and in this
respect, online method of giving input with stylus is gaining quite popularity (Gupta et al., 2007).
Recognition of handwritten characters with respect to any language is difficult due to variability of writing
styles, state of mood of individuals, multiple patterns to represent a single character, cursive representation
of character and number of disconnected and multi-stroke characters (Shanthi and Duraiswamy, 2007).
Current technology supporting pen-based input devices include: Digital Pen by Logitech, Smart Pad by
Pocket PC, Digital Tablets by Wacom and Tablet PC by Compaq (Manuel and Joaquim, 2001). Although
these systems with handwriting recognition capability are already widely available in the market, further
improvements can be made on the recognition performances for these applications.
The challenges posed by the online handwritten character recognition systems are to increase the
recognition accuracy and to reduce the recognition time (Rejean and Sargurl, 2000; Gupta et. al., 2007).
Various approaches that have been used by many researchers to develop character recognition systems,
these include; template matching approach, statistical approach, structural approach, neural networks
approach and hybrid approach. Hybrid approach (combination of multiple classifiers) has become a very
active area of research recently (Kittler and Roli, 2000; 2001). It has been demonstrated in a number of
applications that using more than a single classifier in a recognition task can lead to a significant
improvement of the system’s overall performance. Hence, hybrid approach seems to be a promising
approach to improve the recognition rate and recognition accuracy of current handwriting recognition
systems (Simon and Horst, 2004). However, Selection of a feature extraction method is probably the single
most important factor in achieving high recognition performance in character recognition system (Pradeep
et. al., 2011). No matter how sophisticated the classifiers and learning algorithms, poor feature extraction
will always lead to poor system performance (Marc et. al., 2001). In furtherance, Fenwa et al.(2012)


                                                      90
Computer Engineering and Intelligent Systems                                                   www.iiste.org
ISSN 2222-1719 (Paper) ISSN 2222-2863 (Online)
Vol 3, No.4, 2012

developed a feature extraction technique for online character recognition system using hybrid of
geometrical and statistical features. Thus, through the integration of geometrical and statistical features,
insights were gained into new character properties, since these types of features were considered to be
complementary.


2. Research Methodology
The five stages involved in developing the proposed character recognition system, which include data
acquisition, pre-processing, character processing that comprises feature extraction and character digitization,
training and classification using hybrid neural network model and testing, are as shown in Figure 2.2.
Experiments were performed with 6200 handwriting character samples (English uppercase, lowercase and
digits) collected from 50 subjects using G-Pen 450 digitizer and the system was tested with 100 character
samples written by people who did not participate in the initial data acquisition process. The performance
of the system was evaluated based on different learning rates, different image sizes and different database
sizes.

2.1 Data Acquisition
The data used in this work were collected using Digitizer tablet (G-Pen 450) shown in Figure 2.3. It has an
electric pen with sensing writing board. An interface was developed using C# to acquire data (character
information) such as stroke number, stroke pressure, etc from different subjects using the digitizer tablet. 26
Upper case (A-Z), 26 lower case (a-z) English alphabets and 10 digits (0-9) making a total number of 62
characters. 6,200 characters (62 x 2 x 50) were collected from 50 subjects as each individual was requested
to write each of the characters 2 times (this is done to allow the network learn various possible variations of
a single character and become adaptive in nature). This serves as the training data set which was the input
data that was fed into the neural network.

2.2 Data Preprocessing
Pre-processing is done prior to the application of feature extraction algorithms. Pre-processing aims to
produce clean character images that are easy for the character recognition systems to operate more
accurately. Feature extraction stage relies on the output of this process. The pre processing techniques used
in this research work is Grid Resizing.

2.3 Resizing Grid
From the interface that was provided, there wasn’t any degree of measurement to determine how small/big
the input character should be. Hence, the character written is resized to a matrix size of 5 by 7, 10 by 14
and 20 by 28 for all input characters. This is used to get the universe of discourse which is the shortest
matrix that fit the entire character skeleton. The universe of discourse is measured to easily get a uniform
matrix size in multiple of 5 by 7, 10 by 14 and 20 by 28 respectively.
Any character measured smaller than the required size is considerably resized to the multiple of 5 by 7, 10
by 14 and 20 by 28 conversely any character measured larger than the required size will also be resized to
the multiple of 5 by 7, 10 by 14 and 20 by 28. This implies that rows and columns of Zero’s are added or
subtracted from the resized image matrix to achieve the required multiple of 5 by 7, 10 by 14 and 20 by 28.


2.4 Feature Extraction Development
The goal of feature extraction is to extract a set of features, which maximizes the recognition rate with the
least amount of elements. Many feature extraction techniques have been proposed to improve overall
recognition rate; however, most of the techniques used only one property of the handwritten character. This
research focuses on a feature extraction technique that combined three characteristics (stroke information,
contour pixels and zoning) of the handwritten character to create a global feature vector. Hence, a hybrid
feature extraction algorithm was developed using Geometrical and Statistical features as shown in Figure


                                                      91
Computer Engineering and Intelligent Systems                                                   www.iiste.org
ISSN 2222-1719 (Paper) ISSN 2222-2863 (Online)
Vol 3, No.4, 2012

2.4. Integration of Geometrical and Statistical features was used to highlight different character properties,
since these types of features are considered to be complementary.


2.4.1 The Developed Hybrid (Geom-Statistical) Feature Extraction Algorithm
        The Hybrid feature extraction model adopted in this work is as shown in Figure 2.4. Stages of
development of the proposed hybrid feature extraction algorithm are as follow:
Stage 1: Get the stroke information of the input characters from the digitizer (G- pen 450). These include:
            (i)    Pressure used in writing the strokes of the characters
            (ii)   Number (s) of strokes used in writing the characters
         (iii)     Number of junctions and the location in the written characters
         (iv)      The horizontal projection count of the character.
Stage 2: Apply Contour tracing algorithm to trace out the contour of the characters:


Stage 3: Develop a modified hybrid zoning algorithm and RUN it on the contours of the characters:
         Two zoning algorithms were proposed by Vanajah and Rajashekararadhya in 2008 for the
         recognition of four popular Indian scripts (Kannada, Telugu, Tamil and Malayalam) numeral.
         These were: Image Centroid and Zone-based (ICZ) Distance Metric Feature Extraction System
         (Vanajah and Rajashekararadhya, 2008a) and Zone Centroid and Zone-based (ZCZ) Distance
         Metric Feature Extraction System (Rajashekararadhya and Vanajah, 2008b). The Two Algorithms
         were modified in terms of:
         (i)       Number of zones being used (25 Zones) as shown in Figure 2.1.
         (ii)      Measurement of the distances from both the Image Centroid and Zone Centroid to the
                   pixels present in each zone.
         (iii)     The area of application (uppercase (A-Z), lowercase (a-z) English alphabet and digit (0-
                   9)).
Few zones are adopted in this research but emphasis is laid on how to effectively measure the pixel
densities in each zone. However, pixels pass through the zones at varied distances that is why we measured
five distances for each zone at an angle of 200 .


Hybrid of Modified ICZ and Modified ZCZ based Distance Metric Feature Extraction Algorithm.


Input: Pre processed character image
Output: Features for classification and recognition
Method Begins
Step1: Divide the input image into 25 equal zones.
Step2: Compute the input image centroid
Step3: Compute the distance between the image centroid to each pixel present in the zone measured at
angle 200
Step4: Repeat step 3 for the entire pixel present in the zone.
Step5: Compute average distance between these points.


                                                      92
Computer Engineering and Intelligent Systems                                                    www.iiste.org
 ISSN 2222-1719 (Paper) ISSN 2222-2863 (Online)
 Vol 3, No.4, 2012

 Step6: Compute the zone centroid
 Step7: Compute the distance between the zone centroid to each pixel present in the zone measured at angle
 200.
 Step8: Repeat step 7 for the entire pixel present in the zone
 Step9: Compute average distance between these points.
 Step10: Repeat the steps 3-9 sequentially for the entire zones.
 Step11: Finally, 2xn (50) such features were obtained for classification and recognition.
 Method Ends


Stage 4: Feed the outputs of the extracted features of the characters into the digitization stage in order to
          convert all the extracted features into digital forms.

 2.4.2 Development of Hybrid Neural Network Model
 In this research work, hybrid approach with serial combination scheme was adopted. Hybrid of modified
 Counterpropagation and modified Optical Backpropagation neural network was developed as shown in
 Figure 2.5. The architecture of the neural network with three layers is illustrated in Figure 2.5. The output
 of ith layer is given by (2.1) except the output layer which uses maxsoft function:
 ai = logsig(wi ai-1 + bi)
                   (2.1)
         where i = (2, 3) and a0 = P, a1 = E where E is the Euclidean distance between the weight vector and
 the input vector.
         wi = Weight vector of ith layer
         ai = Output of ith layer
         bi = Bias vector for ith layer.
 The input vector ‘P’ is represented by the solid vertical bar at the left. The dimensions of ‘P’ are displayed
 as 35 x 1, indicating that the input is a single vector of 35 elements (i.e. the image size). These inputs go to
 weight matrix ‘W1’, which has 86 rows (i.e. 86 neurons in the first hidden layer) and 35 columns. A
 constant ‘1’ enters the neuron as input and is multiplied by a bias ‘b1’. The net input to the transfer function
 (Euclidean distance) in the Kohonen (hidden layer) is ‘n1’, which is given as the Euclidean distance
 between the weight vector wi and the input vector P. The neurons output ‘a1’ serves as inputs to the second
 hidden layer. These inputs go to weight matrix ‘W2’, which has 86 rows (i.e.86 neurons in the first hidden
 layer) and 35 columns. A constant ‘1’ enters the neuron as input and is multiplied by a bias ‘b2’.
 The net input to the transfer function (log sigmoid) in the second (hidden layer) is ‘n2’, which is the sum of
 the bias ‘b2’ and the product ‘W2a1’. The output ‘a2’ is a single vector of 86 elements and this serves as
 inputs to the output layer ‘a3’. These inputs go to weight matrix ‘W3’, which has 86 rows (i.e. 86 neurons
 from the second hidden layer) and 62 columns (26 uppercase + 26 lowercase + 10 digits). The net input to
 the transfer function (maxsoft) in the output layer is ‘n3’, which is the sum of the bias ‘b3’ and the product
 ‘W3a2’. The neurons output ‘a3’ is a single vector of 62 elements and this serves as the final output of the
 neural network.


 This research work adopted the modified Counterpropagation network (CPN) developed by Jude, Vijila,
 and Anitha (2010). The training algorithm involves the following two phases:
 (i) Weight adjustment between the input layer and the hidden layer (Kohonen layer)
 The weight adjustment procedure for the hidden layer weights is same as that of the conventional CPN. It
 follows the unsupervised methodology to obtain the stabilized weights. After convergence, the weights
 between the hidden layer and the output layer are calculated.


                                                        93
Computer Engineering and Intelligent Systems                                                   www.iiste.org
ISSN 2222-1719 (Paper) ISSN 2222-2863 (Online)
Vol 3, No.4, 2012

(ii) Weight adjustment between the hidden layer and its output layer
The weight adjustment procedure employed in this work is significantly different from the conventional
CPN. The weights are calculated in the reverse direction without any iterative procedures. Normally, the
weights are calculated based on the criteria of minimizing the error. But in this work, a minimum error
value is specified initially and the weights are estimated based on the error value. Thus without any training
methodology, the weight values are estimated. This technique accounts for higher convergence rate since
one set of weights are estimated directly. It is this output that served as the input to the modified optical
backpropagation algorithm.

2.4.2.1 Modified Optical Backpropagation Neural Network
In the standard backpropagation, the error at a single output unit is defined according to Equation as:
δo pk = (Ypk - Opk ) . fo' k (neto pk)
          (2.2)
where the subscript “p” refers to the pth training vector, and “k” refers to the kth output unit, Ypk is the
desired output value, Opk is the actual output from kth unit, then δopk will propagate backward to update the
output-layer weights and the hidden-layer weights. In the Optical backpropagation (OBP), error at a single
output unit is adjusted according to Otair and Salameh (2005) as:
New δopk = (1+e (Ypk - Opk )2 . fo'k (neto pk)), if (Y – O) >= zero
(2.3a)
New δopk = - (1+e (Ypk - Opk )2 . fo'k (neto pk)),  if (Y – O) < zero
(2.3b)

The error function defined in Optical Backpropagation (Otair and Salameh, 2005) earlier is proportional to
the square of the Euclidean distance between the desired output and the actual output of the network for a
particular input pattern. As an alternative, other error functions whose derivatives exist and can be
calculated at the output layer can replace the traditional square error criterion (Haykin, 2003). In this
research work, error of the third order (Cubic error) has been adopted to replace the traditional square error
criterion used in Optical backpropagation. The Equation of the cubic error is given as:
 δopk = -3(Ypk – Opk)2 . fo'k(netopk)
(2.4)
The cubic error in Equation (2.4) was manipulated mathematically in order to maximize the error of each
output unit which will be transmitted backward from the output layer to each unit in the intermediate layer.
These are as shown in Equations (2.5a) and (2.5b) below:

Modified δopk = 3((1+ et)2 . fo'k(netopk))     If (Ypk – Opk)2 >= 0
(2.5a)
Modified δopk = -3((1+ et)2 . fo'k(netopk))    If (Ypk – Opk)2 <=0
(2.5b)                   where Ypk = Target or Desired output
                        Opk = Network output
                          t = (Ypk – Opk)2
However, one of the ways to reduce the training time is through the use of momentum, as it enhances the
stability of the training process. The momentum is used to keep the training process going in the same
general direction (Haykin, 2003). In the modified Optical Backpropagation network, momentum was
introduced. Hence, the weight update for the output unit is:
Wokj(t+1) = Wokj(t) + µWokj(t) + (η . Modified δopk. ipj)
(2.6)
where µ is the momentum coefficient typically about 0.9 and η is the learning rate.

2.4.2.1 The Modified Optical Backpropagation Algorithm
Modifications of the algorithm are in terms of:



                                                     94
Computer Engineering and Intelligent Systems                                                   www.iiste.org
ISSN 2222-1719 (Paper) ISSN 2222-2863 (Online)
Vol 3, No.4, 2012

(i)    Error signal function
(ii)   Application area

With the introduction of Cubic error function and Momentum, the modified Optical Backpropagation is
given as:
1.       Apply the input example to the input units.
2.       Calculate the net-input values to the hidden layer units.
3.       Calculate the outputs from the hidden layer.
4.       Calculate the net-input values to the output layer units
5.       Calculate the outputs from the output units
6.       Calculate the error term for the output units, using Equations (2.5a) and (2.5b)
7.       Calculate the error term for the hidden units, through applying Modified δopk also
         Modified δh pj = fh'j(neth pj) .(              δopk . Wokj)
         (2.7)
8.       Update weights on the output layer.
         Wokj(t+1) = Wokj(t) + µWokj(t) + (η . Modified δopk . ipj)
         (2.8)
9.       Update weights on the hidden layer.
         Whji(t+1) = Whji(t) + (η . Modified δhpj . Xi)
                        (2.9)
Repeat steps from step 1 to step 9 until the error (Ypk – Opk) is acceptably small for each of the training
vector pair. The proposed algorithm as in OBP is stopped when the cubes of the differences between the
actual and target values summed over units and all patterns are acceptably small.

2.4.3 The Hybrid Neural Network Algorithm
This research work employed hybrid of modified Counterpropagation and modified Optical
Backpropagation neural networks for the training and classification of the input pattern. The training
algorithm involves the following two stages:
Stage A: Performs the training of the weights from the input nodes to the Kohonen hidden node.
Step 1: Weight adjustment between the input layer and the hidden layer

The weight adjustment procedure for the hidden layer weights is same as that of the conventional CPN. It
follows the unsupervised methodology to obtain the stabilized weights. This process is repeated for a
suitable number of iterations and the stabilized set of weights are obtained. After convergence, the weights
between the Kohonen hidden layer and the output layer are calculated.

Step 2: Weight adjustment between the hidden layer and the output layer

         The weight adjustment procedure employed in this work is significantly different from the
conventional CPN. The weights are calculated in the reverse direction without any iterative procedures.
Normally, the weights are calculated based on the criteria of minimizing the error. But in this work, a
minimum error value is specified initially and the weights are estimated based on the error value. The
detailed steps of the modified algorithm are given below.

Step 1: The stabilized weight values are obtained when the error value (target output) is equal to zero (or) a
predefined minimum value. The error value used for convergence in this work is 0.1. The following
procedure uses this concept for weight matrices calculation.

Step 2: Supply the target vectors t1 to the output layer neurons

Step 3: Since ( t1 – y1 ) = 0.1 for convergence
(2.10)

                                                     95
Computer Engineering and Intelligent Systems                                                     www.iiste.org
ISSN 2222-1719 (Paper) ISSN 2222-2863 (Online)
Vol 3, No.4, 2012

The output of the output layer neurons is set equal to the target values as:

          y1 = t1 − 0.1
(2.11)

Step 4: Once the output value is calculated, the sum of the weighted input signals. Thus without any
training methodology, the weight values are estimated. This technique accounts for higher convergence rate
since one set of weights are estimated directly.

STAGE B: Performs the training of the weights from the second hidden node to the output nodes.
1.       Calculate the net-input values from the Kohonen to the second hidden layer units.
2.       Calculate the outputs from the second hidden layer.
3.       Calculate the net-input values to the output layer units
4.       Calculate the outputs from the output units
5.       Calculate the error term for the output units, using Equations (2.5a) and (2.5b)
6.       Calculate the error term for the hidden units, through applying modified δopk as in Equation (2.7)
7.       Update weights on the output layer using (2.8)
8.       Update weights on the hidden layer using Equation (2.9)
Repeat steps from step 1 to step 8 until the error (Ypk – Opk) is acceptably small for each of the training
vector pair. The proposed algorithm as classical BP is stopped when the cubes of the differences between
the actual and target values summed over units and all patterns are acceptably small.

         The description of notation used in the training procedure is as given below:

         Xpi :     net input to the ith input unit

         nethpj:          net input to the jth hidden unit

         Whji:            weight on the connection from the ith input unit to jth hidden unit

         ipj:             net input to the jth hidden unit

         netopk:          net input to the kth output unit

         Wokj:            weight on the connection from the jth hidden unit to kth output unit

         Opk:             actual output for the kth output unit



3. Software Implementation
This section discussed the phases of developing the proposed character recognition system. All the
algorithms have been implemented using C# programming language and RUN under Windows7 operating
system on Pentium (R) 2.00GB RAM, 1.83GH Processor and Pentium (R) 4.00GB RAM, 2.13GH
Processor respectively. Different interfaces representing the phases of the system development are shown
from Figures 2.6 to 2.11.

3.1 Character Acquisition
Character acquisition interface as in Figures 2.7 and 2.8 displayed a drawing area to serve as a platform to
acquire characters from users. Any drawn character on this platform must be saved into the developed
system’s database by clicking the ‘Add Botton’. It also captured number of strokes and pressure used in
writing a character. A total 6,200 character samples were acquired using G-Pen 450 as shown in Figure 2.2

                                                         96
Computer Engineering and Intelligent Systems                                                  www.iiste.org
ISSN 2222-1719 (Paper) ISSN 2222-2863 (Online)
Vol 3, No.4, 2012

and stored in the database.

3.2 Network Training
Before the network can be trained, the network parameters such as the learning rate, epoch value, the quit
error and character image size must be specified in the ‘Application setting’ interface. The next thing is to
specify the total number of the character to be loaded from the database for training. The database size can
be varied by specifying the maximum number of character to be selected from character category
(uppercase, lowercase and digit). This can be accomplished with the interfaces shown in Figure 2.9 and
Figure 2.10. On clicking the ‘Training phase button’, the neural network trains itself with alphanumeric
characters of A – Z, a – z and 0 – 9 by performing several iterations based on the error value, learning rate
and the epoch value stipulated for the Neural Network. Epoch increases from 1 to 10,000 and the Error
reduces from a particular value depending on the value specified by the user. The training is terminated
when either the epoch reaches 10,000 or the error reaches the minimum value specified. Network training
results such the training time, the epoch value were displayed as shown in Figure 2.10.


3.3 Recognition Phase
Recognition of the character can be accomplished by clicking the ‘Data acquisition/ Recognition phase
botton’. User is required to write a character and click the ‘Recognize botton’. The results such as the
‘detected character’, ‘recognition rate’ and ‘recognition time’ will be displayed as shown in Figure 2.11.


4. Performance Evaluation
The performance metrics adopted in evaluating the developed system are: different learning rate parameters,
different image sizes, different database sizes and system configurations. The results are as given in Figure
2.12 to Figure 2.17. Figure 2.12 shows the graph of learning rate versus epoch. Learning rate parameter
variation has a positive effect on the network performance. The smaller the value of learning rate, the lower
the value with which the network updates its weights. This intuitively implies that it will be less likely to
face the over learning difficulty since it will be updating its links slowly and in a more refined manner.
However, this would also imply more number of iterations is required to reach its optimal state. Figure 2.13
indicates the graph of image size versus epoch. The image size is directly proportional to epoch. Usually,
the complex and large sized input sets require a large topology network with more number of iterations and
this has an adverse effect on the recognition time. Figure 2.14 shows the graph of variation of database size
and recognition rates. Increase in database size has a positive proportionality relation to the recognition
rates due to the fact that network is able to attribute the test character to larger character samples in the
vector space. However, rate of increment in recognition rate with respect to database size is considerably
small. Moreover, it can be shown from Figure 2.15 that the more the dimensional input vector (image
matrix size), the less the recognition performance. Three different image sizes (5 by 7, 10 by 14 and 20 by
28) were considered; it was shown from the results that the higher the image size, the higher the percentage
of recognition, although, the rate of change was small. Furthermore, increase in image size also increases
the recognition time.

System configuration is also another factor influencing the performance of the recognition system. Both the
training time and recognition time are measured in CPU (processor) seconds, the higher the speed of the
system (processor speed), the lower the training time and the recognition time. This is due to the fact that
more time will be used by 1.8GH system to reach the required epochs (iteration) than the system with
2.13GH speed. The result is as shown in Figure 2.16. Figure 2.17 is the graph showing comparison of
performance of the developed system with related works in literature. It can be shown from the graph that
the accuracy of the system developed by Muhammad et. al. (2005) is higher than that of Muhammad et. al.
(2006). This shows that the performance of Counterpropagation neural network is better than that of
Standard Backpropagation. The best recognition rate was achieved in the developed system with no


                                                     97
Computer Engineering and Intelligent Systems                                                  www.iiste.org
ISSN 2222-1719 (Paper) ISSN 2222-2863 (Online)
Vol 3, No.4, 2012

recognition failure. This means that the developed system was able to recognise characters irrespective of
the writing styles.



5. Conclusion and Future Work
In this paper, we have been able to develop an effective feature extraction technique for the proposed online
character recognition system using hybrid of geometrical and statistical features. The hybrid feature
extraction was developed to alleviate the problem of poor feature extraction algorithm of online character
recognition system. However, a hybrid of modified counter propagation and modified optical
backpropagation neural network model was developed for the proposed system. The performance of the
online character recognition system under consideration was evaluated based on different learning rates,
image sizes and database sizes. The results from the study were compared with some works in the literature
using counterpropagation and standard backpropagation respectively. The results showed that
counterpropagation performed better than standard backpropagation in terms of correct recognition, false
recognition, recognition failure and the achieved recognition rates were 94% and 81% respectively. The
developed system achieved better performance with no recognition failure, 99% recognition rate and an
average recognition time of 2 milliseconds. Future work could be geared towards integrating an
optimization algorithm into the learning algorithms to further enhance the convergence of the neural
network.


References
Anoop, M. N. and Anil K.J. (2004): “Online Handwritten Script Recognition’’, IEEE Trans. PAMI, 26(1):
124
130.
Liu, C.L., Nakashima, K., Sako, H. and Fujisawa, H. (2004): “Handwritten Digit Recognition:
Investigation of
Normalization and Feature Extraction Techniques’’, Pattern Recognition, 37(2): 265-279.
Mohamad, D. and Zafar, M.F. (2004): “Comparative Study of Two Novel Feature Vectors for Complex
Image
Matching Using Counterpropagation Neural Network’’, Journal of Information Technology, FSKSM, UTM,
16(1): 2073-2081.
Naser, M.A., Adnan, M., Arefin, T.M., Golam, S.M. and Naushad, A. (2009): “Comparative Analysis of
Radon and Fan-beam based Feature Extraction Techniques for Bangla Character Recognition”, IJCSNS
International Journal of Computer Science and Network Security, 9(9): 120-135.
Pradeep, J., Srinivasan, E. and Himavathi, S. (2011): “Diagonal Based Feature Extraction for Handwritten
Alphabets Recognition using Neural Network’’, International Journal of Computer Science and Information
Technology (IJCS11), 3(1): 27-37.
Shanthi, N., and Duraiwamy, K. (2007): “Performance Comparison of Different Image Sizes for
Recognizing Unconstrained Handwritten Tamil Characters Using SVM’’, Journal of Science, 3(9): 760-764.


Santosh, K.C. and Nattee, C. (2009): “A Comprehensive Survey on Online Handwriting Recognition
Technology and Its Real Application to the Nepalese Natural Handwriting’’, Kathmandu University Journal
of Science, Engineering Technology, 5(1): 31-55.
Simon G. and Horst B. (2004): “Feature Selection Algorithms for the Generalization of Multiple Classifier
Systems and their Application to Handwritten Word Recognition’’, Pattern Recognition Letters, 25(11):
1323-1336.

                                                     98
Computer Engineering and Intelligent Systems                                                    www.iiste.org
ISSN 2222-1719 (Paper) ISSN 2222-2863 (Online)
Vol 3, No.4, 2012

Gupta, K., Rao, S.V., and Viswanath (2007): “Speeding up Online Character Recognition’’, Proceedings of
Image and Vision Computing Newzealand, Hamilton: 41-45.
Manuel, J., Fonseca, and Joaquim, A.J., (2001): “Experimental Evolution of an Online Scribble
Recognizer’’, Pattern Recognition Letters, 22(12): 1311-1319.
Rejean, P. and Sargur, S.N. (2000): “On-line and Off-line Recognition: A Comprehensive Survey’’, IEEE
Transaction on Pattern Analysis and Machine Intelligence, 22(1): 63-84.
Kittler, J. and Roli, F. (2000): “1st International Workshop on Multiple Classifier Systems”, Cagliari, Italy.
Kittler, J. and Roli, F. (2001): “2nd International Workshop on Multiple Classifier Systems”, Cagliari, Italy.
Marc, P., Alexandre, L., and Christian, G. (2001): “Character Recognition Experiment using Unipen Data’’,
Parizeau and al., Proceeding of ICDAR, Seattle: 10-13.
Muhammad, F.Z., Dzuulkifli, M., and Razib, M.O. (2006): “Writer Independent Online Handwritten
Character Recognition Using Simple Approach’’, Information Technology Journal, 5(3): 476-484.
Freeman, J.A. and Skapura, D.M. (1992): “Backpropagation Neural Networks Algorithm Applications and
Programming Techniques”, Addison-Wesley Publishing Company: 89-125.
Minai, A.A. and Williams, R.D. (1990): “Acceleration of Backpropagation through learning Rate
Momentum Adaptation’’, Proceeding of the International Joint Conference on Neural Networks: 1676-1679.
Riedmiller, M. and Braun, H. (1993): “A Direct Adaptive Method for Faster Backpropagation learning the
PROP Algorithm’’, Proceedings of the IEEE International Conference on Neural Networks (ICNN), 1: 586-
591, Francisco.
Otair, M.A. and Salameh, W.A. (2005a): “Online Handwritten Character Recognition using an Optical
Backpropagation Neural Network’’, Issues in Informing Science and Information Technology, 2: 787-797.
Rajashekararadhya, S.V. and Vanaja, P.R. (2008a): “Handwritten numeral recognition of three popular
South Indian Scripts: A Novel Approach:’’, Proceedings of the Second International Conference on
information processing ICIP: 162-167.
Rajashekararadhya S.V. and Vanaja, P.R (2008b): “Isolated Handwritten Kannada Digit Recognition: A
Novel Approach’’, Proceedings of the International Conference on Cognition and Recognition: 134-140.

Fenwa, O. D., Omidiora, E. O. and Fakolujo O. A. (2012): “Development of a Feature Extraction

Technique for Online Character Recognition System, Journal of Innovative Systems Design and

Engineering, International Institute of Science, Technology and Education,Vol. 3, No.3, pp. 10-23.




                                                      99
Computer Engineering and Intelligent Systems                                             www.iiste.org
ISSN 2222-1719 (Paper) ISSN 2222-2863 (Online)
Vol 3, No.4, 2012

                              Figure 2.1: Character ‘n’ in 5 by 5 (25 equal zones)




             Figure 2.2: Block Diagram of the Developed Character Recognition System




        Figure 2.3: The snapshot of Genius Pen (G-Pen 450) Digitizer for character acquisition




                                                 100
Computer Engineering and Intelligent Systems                                         www.iiste.org
ISSN 2222-1719 (Paper) ISSN 2222-2863 (Online)
Vol 3, No.4, 2012




                    Figure 2.4: The Developed Hybrid Feature Extraction Model




                  a1 = E (W1 – P), a2 = logsig (W2a1 + b2), a3 = Maxsoft (W3 + b3)


                          Figure 2.5: The Hybrid Neural Network Model




                                                101
Computer Engineering and Intelligent Systems                                            www.iiste.org
ISSN 2222-1719 (Paper) ISSN 2222-2863 (Online)
Vol 3, No.4, 2012




       Figure 2.6: Graphic user interface of the developed Online Character Recognition System




                       Figure 2.7: Acquisition of character ’A’ using 3 strokes




                                                 102
Computer Engineering and Intelligent Systems                                              www.iiste.org
ISSN 2222-1719 (Paper) ISSN 2222-2863 (Online)
Vol 3, No.4, 2012


                       Figure 2.8: Acquisition of character ’A’ using 2 strokes




        Figure 2.9: Network parameter setting and loading data to be trained from the database




                              Figure 2.10: Network Training in progress




               Figure 2.11: Result of recognition process displaying recognition status




                                                 103
Computer Engineering and Intelligent Systems                                             www.iiste.org
ISSN 2222-1719 (Paper) ISSN 2222-2863 (Online)
Vol 3, No.4, 2012




    Figure 2.12: Graph showing the effect of variation in Learning Rate and Database size on Epochs




      Figure 2.13: Graph showing the effect of variation of Image size and Database size on Epoch




 Figure 2.14: Graph showing the effect of variation in Database size and Epoch on the Recognition Rate




                                                 104
Computer Engineering and Intelligent Systems                                               www.iiste.org
ISSN 2222-1719 (Paper) ISSN 2222-2863 (Online)
Vol 3, No.4, 2012




Figure 2.15: Graph showing the effect of variation in Image size and Database size on the Recognition Rate




 Figure 2.16: Graph showing the effect of variation in system configurations and Database size on Epochs




                  CR = Correct Recognition; FR = False Recognition; RF = Recognition Failure


     Figure 2.17: Graph showing performance evaluation of the developed system with related works




                                                   105
International Journals Call for Paper
The IISTE, a U.S. publisher, is currently hosting the academic journals listed below. The peer review process of the following journals
usually takes LESS THAN 14 business days and IISTE usually publishes a qualified article within 30 days. Authors should
send their full paper to the following email address. More information can be found in the IISTE website : www.iiste.org

Business, Economics, Finance and Management               PAPER SUBMISSION EMAIL
European Journal of Business and Management               EJBM@iiste.org
Research Journal of Finance and Accounting                RJFA@iiste.org
Journal of Economics and Sustainable Development          JESD@iiste.org
Information and Knowledge Management                      IKM@iiste.org
Developing Country Studies                                DCS@iiste.org
Industrial Engineering Letters                            IEL@iiste.org


Physical Sciences, Mathematics and Chemistry              PAPER SUBMISSION EMAIL
Journal of Natural Sciences Research                      JNSR@iiste.org
Chemistry and Materials Research                          CMR@iiste.org
Mathematical Theory and Modeling                          MTM@iiste.org
Advances in Physics Theories and Applications             APTA@iiste.org
Chemical and Process Engineering Research                 CPER@iiste.org


Engineering, Technology and Systems                       PAPER SUBMISSION EMAIL
Computer Engineering and Intelligent Systems              CEIS@iiste.org
Innovative Systems Design and Engineering                 ISDE@iiste.org
Journal of Energy Technologies and Policy                 JETP@iiste.org
Information and Knowledge Management                      IKM@iiste.org
Control Theory and Informatics                            CTI@iiste.org
Journal of Information Engineering and Applications       JIEA@iiste.org
Industrial Engineering Letters                            IEL@iiste.org
Network and Complex Systems                               NCS@iiste.org


Environment, Civil, Materials Sciences                    PAPER SUBMISSION EMAIL
Journal of Environment and Earth Science                  JEES@iiste.org
Civil and Environmental Research                          CER@iiste.org
Journal of Natural Sciences Research                      JNSR@iiste.org
Civil and Environmental Research                          CER@iiste.org


Life Science, Food and Medical Sciences                   PAPER SUBMISSION EMAIL
Journal of Natural Sciences Research                      JNSR@iiste.org
Journal of Biology, Agriculture and Healthcare            JBAH@iiste.org
Food Science and Quality Management                       FSQM@iiste.org
Chemistry and Materials Research                          CMR@iiste.org


Education, and other Social Sciences                      PAPER SUBMISSION EMAIL
Journal of Education and Practice                         JEP@iiste.org
Journal of Law, Policy and Globalization                  JLPG@iiste.org                       Global knowledge sharing:
New Media and Mass Communication                          NMMC@iiste.org                       EBSCO, Index Copernicus, Ulrich's
Journal of Energy Technologies and Policy                 JETP@iiste.org                       Periodicals Directory, JournalTOCS, PKP
Historical Research Letter                                HRL@iiste.org                        Open Archives Harvester, Bielefeld
                                                                                               Academic Search Engine, Elektronische
Public Policy and Administration Research                 PPAR@iiste.org                       Zeitschriftenbibliothek EZB, Open J-Gate,
International Affairs and Global Strategy                 IAGS@iiste.org                       OCLC WorldCat, Universe Digtial Library ,
Research on Humanities and Social Sciences                RHSS@iiste.org                       NewJour, Google Scholar.

Developing Country Studies                                DCS@iiste.org                        IISTE is member of CrossRef. All journals
Arts and Design Studies                                   ADS@iiste.org                        have high IC Impact Factor Values (ICV).

Contenu connexe

Tendances

Volume 2-issue-6-2009-2015
Volume 2-issue-6-2009-2015Volume 2-issue-6-2009-2015
Volume 2-issue-6-2009-2015
Editor IJARCET
 
Bratislava WS - Fuchs - Abbyy - OCR overview_pdf
Bratislava WS - Fuchs - Abbyy - OCR overview_pdfBratislava WS - Fuchs - Abbyy - OCR overview_pdf
Bratislava WS - Fuchs - Abbyy - OCR overview_pdf
IMPACT Centre of Competence
 
CONTENT RECOVERY AND IMAGE RETRIVAL IN IMAGE DATABASE CONTENT RETRIVING IN TE...
CONTENT RECOVERY AND IMAGE RETRIVAL IN IMAGE DATABASE CONTENT RETRIVING IN TE...CONTENT RECOVERY AND IMAGE RETRIVAL IN IMAGE DATABASE CONTENT RETRIVING IN TE...
CONTENT RECOVERY AND IMAGE RETRIVAL IN IMAGE DATABASE CONTENT RETRIVING IN TE...
Editor IJMTER
 

Tendances (19)

Optical character recognition IEEE Paper Study
Optical character recognition IEEE Paper StudyOptical character recognition IEEE Paper Study
Optical character recognition IEEE Paper Study
 
Volume 2-issue-6-2009-2015
Volume 2-issue-6-2009-2015Volume 2-issue-6-2009-2015
Volume 2-issue-6-2009-2015
 
Basics of-optical-character-recognition
Basics of-optical-character-recognitionBasics of-optical-character-recognition
Basics of-optical-character-recognition
 
Ocr abstract
Ocr abstractOcr abstract
Ocr abstract
 
CRC Final Report
CRC Final ReportCRC Final Report
CRC Final Report
 
Optical Character Recognition (OCR) System
Optical Character Recognition (OCR) SystemOptical Character Recognition (OCR) System
Optical Character Recognition (OCR) System
 
OCR Text Extraction
OCR Text ExtractionOCR Text Extraction
OCR Text Extraction
 
Optical Character Recognition Using Python
Optical Character Recognition Using PythonOptical Character Recognition Using Python
Optical Character Recognition Using Python
 
Optical Character Recognition (OCR) based Retrieval
Optical Character Recognition (OCR) based RetrievalOptical Character Recognition (OCR) based Retrieval
Optical Character Recognition (OCR) based Retrieval
 
ocr
ocrocr
ocr
 
Bratislava WS - Fuchs - Abbyy - OCR overview_pdf
Bratislava WS - Fuchs - Abbyy - OCR overview_pdfBratislava WS - Fuchs - Abbyy - OCR overview_pdf
Bratislava WS - Fuchs - Abbyy - OCR overview_pdf
 
Optical Character Recognition( OCR )
Optical Character Recognition( OCR )Optical Character Recognition( OCR )
Optical Character Recognition( OCR )
 
Optical Character Recognition (OCR)
Optical Character Recognition (OCR)Optical Character Recognition (OCR)
Optical Character Recognition (OCR)
 
OCR2
OCR2OCR2
OCR2
 
Optical Character Recognition
Optical Character RecognitionOptical Character Recognition
Optical Character Recognition
 
CONTENT RECOVERY AND IMAGE RETRIVAL IN IMAGE DATABASE CONTENT RETRIVING IN TE...
CONTENT RECOVERY AND IMAGE RETRIVAL IN IMAGE DATABASE CONTENT RETRIVING IN TE...CONTENT RECOVERY AND IMAGE RETRIVAL IN IMAGE DATABASE CONTENT RETRIVING IN TE...
CONTENT RECOVERY AND IMAGE RETRIVAL IN IMAGE DATABASE CONTENT RETRIVING IN TE...
 
OCR (Optical Character Recognition)
OCR (Optical Character Recognition) OCR (Optical Character Recognition)
OCR (Optical Character Recognition)
 
Final Report on Optical Character Recognition
Final Report on Optical Character Recognition Final Report on Optical Character Recognition
Final Report on Optical Character Recognition
 
OCR
OCROCR
OCR
 

Similaire à 11.development of a writer independent online handwritten character recognition system using modified hybrid

Comparative Analysis of PSO and GA in Geom-Statistical Character Features Sel...
Comparative Analysis of PSO and GA in Geom-Statistical Character Features Sel...Comparative Analysis of PSO and GA in Geom-Statistical Character Features Sel...
Comparative Analysis of PSO and GA in Geom-Statistical Character Features Sel...
IJERA Editor
 
Volume 2-issue-6-2009-2015
Volume 2-issue-6-2009-2015Volume 2-issue-6-2009-2015
Volume 2-issue-6-2009-2015
Editor IJARCET
 
Lei_Resume-it.doc
Lei_Resume-it.docLei_Resume-it.doc
Lei_Resume-it.doc
butest
 
Implementation and Performance Evaluation of Neural Network for English Alpha...
Implementation and Performance Evaluation of Neural Network for English Alpha...Implementation and Performance Evaluation of Neural Network for English Alpha...
Implementation and Performance Evaluation of Neural Network for English Alpha...
ijtsrd
 
An offline signature verification using pixels intensity levels
An offline signature verification using pixels intensity levelsAn offline signature verification using pixels intensity levels
An offline signature verification using pixels intensity levels
Salam Shah
 

Similaire à 11.development of a writer independent online handwritten character recognition system using modified hybrid (20)

Using optimized features for modified optical backpropagation
Using optimized features for modified optical backpropagationUsing optimized features for modified optical backpropagation
Using optimized features for modified optical backpropagation
 
Comparative Analysis of PSO and GA in Geom-Statistical Character Features Sel...
Comparative Analysis of PSO and GA in Geom-Statistical Character Features Sel...Comparative Analysis of PSO and GA in Geom-Statistical Character Features Sel...
Comparative Analysis of PSO and GA in Geom-Statistical Character Features Sel...
 
Volume 2-issue-6-2009-2015
Volume 2-issue-6-2009-2015Volume 2-issue-6-2009-2015
Volume 2-issue-6-2009-2015
 
HandwrittenDigitRecognitionusing.pdf
HandwrittenDigitRecognitionusing.pdfHandwrittenDigitRecognitionusing.pdf
HandwrittenDigitRecognitionusing.pdf
 
Handwritten Digit Recognition Using CNN
Handwritten Digit Recognition Using CNNHandwritten Digit Recognition Using CNN
Handwritten Digit Recognition Using CNN
 
D017222226
D017222226D017222226
D017222226
 
IRJET- Intelligent Character Recognition of Handwritten Characters
IRJET- Intelligent Character Recognition of Handwritten CharactersIRJET- Intelligent Character Recognition of Handwritten Characters
IRJET- Intelligent Character Recognition of Handwritten Characters
 
Character Recognition using Data Mining Technique (Artificial Neural Network)
Character Recognition using Data Mining Technique (Artificial Neural Network)Character Recognition using Data Mining Technique (Artificial Neural Network)
Character Recognition using Data Mining Technique (Artificial Neural Network)
 
Telugu letters dataset and parallel deep convolutional neural network with a...
Telugu letters dataset and parallel deep convolutional neural  network with a...Telugu letters dataset and parallel deep convolutional neural  network with a...
Telugu letters dataset and parallel deep convolutional neural network with a...
 
Online Hand Written Character Recognition
Online Hand Written Character RecognitionOnline Hand Written Character Recognition
Online Hand Written Character Recognition
 
A Deep Learning Approach to Recognize Cursive Handwriting
A Deep Learning Approach to Recognize Cursive HandwritingA Deep Learning Approach to Recognize Cursive Handwriting
A Deep Learning Approach to Recognize Cursive Handwriting
 
IRJET- Offline Transcription using AI
IRJET-  	  Offline Transcription using AIIRJET-  	  Offline Transcription using AI
IRJET- Offline Transcription using AI
 
Lei_Resume-it.doc
Lei_Resume-it.docLei_Resume-it.doc
Lei_Resume-it.doc
 
O45018291
O45018291O45018291
O45018291
 
Bangla_handwritten_dig1] final proposal .pdf
Bangla_handwritten_dig1] final proposal .pdfBangla_handwritten_dig1] final proposal .pdf
Bangla_handwritten_dig1] final proposal .pdf
 
Implementation and Performance Evaluation of Neural Network for English Alpha...
Implementation and Performance Evaluation of Neural Network for English Alpha...Implementation and Performance Evaluation of Neural Network for English Alpha...
Implementation and Performance Evaluation of Neural Network for English Alpha...
 
­­­­Cursive Handwriting Recognition System using Feature Extraction and Artif...
­­­­Cursive Handwriting Recognition System using Feature Extraction and Artif...­­­­Cursive Handwriting Recognition System using Feature Extraction and Artif...
­­­­Cursive Handwriting Recognition System using Feature Extraction and Artif...
 
A Novel Framework For Numerical Character Recognition With Zoning Distance Fe...
A Novel Framework For Numerical Character Recognition With Zoning Distance Fe...A Novel Framework For Numerical Character Recognition With Zoning Distance Fe...
A Novel Framework For Numerical Character Recognition With Zoning Distance Fe...
 
An offline signature verification using pixels intensity levels
An offline signature verification using pixels intensity levelsAn offline signature verification using pixels intensity levels
An offline signature verification using pixels intensity levels
 
Online Handwriting Recognition using HMM
Online Handwriting Recognition using HMMOnline Handwriting Recognition using HMM
Online Handwriting Recognition using HMM
 

Plus de Alexander Decker

Abnormalities of hormones and inflammatory cytokines in women affected with p...
Abnormalities of hormones and inflammatory cytokines in women affected with p...Abnormalities of hormones and inflammatory cytokines in women affected with p...
Abnormalities of hormones and inflammatory cytokines in women affected with p...
Alexander Decker
 
A usability evaluation framework for b2 c e commerce websites
A usability evaluation framework for b2 c e commerce websitesA usability evaluation framework for b2 c e commerce websites
A usability evaluation framework for b2 c e commerce websites
Alexander Decker
 
A universal model for managing the marketing executives in nigerian banks
A universal model for managing the marketing executives in nigerian banksA universal model for managing the marketing executives in nigerian banks
A universal model for managing the marketing executives in nigerian banks
Alexander Decker
 
A unique common fixed point theorems in generalized d
A unique common fixed point theorems in generalized dA unique common fixed point theorems in generalized d
A unique common fixed point theorems in generalized d
Alexander Decker
 
A trends of salmonella and antibiotic resistance
A trends of salmonella and antibiotic resistanceA trends of salmonella and antibiotic resistance
A trends of salmonella and antibiotic resistance
Alexander Decker
 
A transformational generative approach towards understanding al-istifham
A transformational  generative approach towards understanding al-istifhamA transformational  generative approach towards understanding al-istifham
A transformational generative approach towards understanding al-istifham
Alexander Decker
 
A time series analysis of the determinants of savings in namibia
A time series analysis of the determinants of savings in namibiaA time series analysis of the determinants of savings in namibia
A time series analysis of the determinants of savings in namibia
Alexander Decker
 
A therapy for physical and mental fitness of school children
A therapy for physical and mental fitness of school childrenA therapy for physical and mental fitness of school children
A therapy for physical and mental fitness of school children
Alexander Decker
 
A theory of efficiency for managing the marketing executives in nigerian banks
A theory of efficiency for managing the marketing executives in nigerian banksA theory of efficiency for managing the marketing executives in nigerian banks
A theory of efficiency for managing the marketing executives in nigerian banks
Alexander Decker
 
A systematic evaluation of link budget for
A systematic evaluation of link budget forA systematic evaluation of link budget for
A systematic evaluation of link budget for
Alexander Decker
 
A synthetic review of contraceptive supplies in punjab
A synthetic review of contraceptive supplies in punjabA synthetic review of contraceptive supplies in punjab
A synthetic review of contraceptive supplies in punjab
Alexander Decker
 
A synthesis of taylor’s and fayol’s management approaches for managing market...
A synthesis of taylor’s and fayol’s management approaches for managing market...A synthesis of taylor’s and fayol’s management approaches for managing market...
A synthesis of taylor’s and fayol’s management approaches for managing market...
Alexander Decker
 
A survey paper on sequence pattern mining with incremental
A survey paper on sequence pattern mining with incrementalA survey paper on sequence pattern mining with incremental
A survey paper on sequence pattern mining with incremental
Alexander Decker
 
A survey on live virtual machine migrations and its techniques
A survey on live virtual machine migrations and its techniquesA survey on live virtual machine migrations and its techniques
A survey on live virtual machine migrations and its techniques
Alexander Decker
 
A survey on data mining and analysis in hadoop and mongo db
A survey on data mining and analysis in hadoop and mongo dbA survey on data mining and analysis in hadoop and mongo db
A survey on data mining and analysis in hadoop and mongo db
Alexander Decker
 
A survey on challenges to the media cloud
A survey on challenges to the media cloudA survey on challenges to the media cloud
A survey on challenges to the media cloud
Alexander Decker
 
A survey of provenance leveraged
A survey of provenance leveragedA survey of provenance leveraged
A survey of provenance leveraged
Alexander Decker
 
A survey of private equity investments in kenya
A survey of private equity investments in kenyaA survey of private equity investments in kenya
A survey of private equity investments in kenya
Alexander Decker
 
A study to measures the financial health of
A study to measures the financial health ofA study to measures the financial health of
A study to measures the financial health of
Alexander Decker
 

Plus de Alexander Decker (20)

Abnormalities of hormones and inflammatory cytokines in women affected with p...
Abnormalities of hormones and inflammatory cytokines in women affected with p...Abnormalities of hormones and inflammatory cytokines in women affected with p...
Abnormalities of hormones and inflammatory cytokines in women affected with p...
 
A validation of the adverse childhood experiences scale in
A validation of the adverse childhood experiences scale inA validation of the adverse childhood experiences scale in
A validation of the adverse childhood experiences scale in
 
A usability evaluation framework for b2 c e commerce websites
A usability evaluation framework for b2 c e commerce websitesA usability evaluation framework for b2 c e commerce websites
A usability evaluation framework for b2 c e commerce websites
 
A universal model for managing the marketing executives in nigerian banks
A universal model for managing the marketing executives in nigerian banksA universal model for managing the marketing executives in nigerian banks
A universal model for managing the marketing executives in nigerian banks
 
A unique common fixed point theorems in generalized d
A unique common fixed point theorems in generalized dA unique common fixed point theorems in generalized d
A unique common fixed point theorems in generalized d
 
A trends of salmonella and antibiotic resistance
A trends of salmonella and antibiotic resistanceA trends of salmonella and antibiotic resistance
A trends of salmonella and antibiotic resistance
 
A transformational generative approach towards understanding al-istifham
A transformational  generative approach towards understanding al-istifhamA transformational  generative approach towards understanding al-istifham
A transformational generative approach towards understanding al-istifham
 
A time series analysis of the determinants of savings in namibia
A time series analysis of the determinants of savings in namibiaA time series analysis of the determinants of savings in namibia
A time series analysis of the determinants of savings in namibia
 
A therapy for physical and mental fitness of school children
A therapy for physical and mental fitness of school childrenA therapy for physical and mental fitness of school children
A therapy for physical and mental fitness of school children
 
A theory of efficiency for managing the marketing executives in nigerian banks
A theory of efficiency for managing the marketing executives in nigerian banksA theory of efficiency for managing the marketing executives in nigerian banks
A theory of efficiency for managing the marketing executives in nigerian banks
 
A systematic evaluation of link budget for
A systematic evaluation of link budget forA systematic evaluation of link budget for
A systematic evaluation of link budget for
 
A synthetic review of contraceptive supplies in punjab
A synthetic review of contraceptive supplies in punjabA synthetic review of contraceptive supplies in punjab
A synthetic review of contraceptive supplies in punjab
 
A synthesis of taylor’s and fayol’s management approaches for managing market...
A synthesis of taylor’s and fayol’s management approaches for managing market...A synthesis of taylor’s and fayol’s management approaches for managing market...
A synthesis of taylor’s and fayol’s management approaches for managing market...
 
A survey paper on sequence pattern mining with incremental
A survey paper on sequence pattern mining with incrementalA survey paper on sequence pattern mining with incremental
A survey paper on sequence pattern mining with incremental
 
A survey on live virtual machine migrations and its techniques
A survey on live virtual machine migrations and its techniquesA survey on live virtual machine migrations and its techniques
A survey on live virtual machine migrations and its techniques
 
A survey on data mining and analysis in hadoop and mongo db
A survey on data mining and analysis in hadoop and mongo dbA survey on data mining and analysis in hadoop and mongo db
A survey on data mining and analysis in hadoop and mongo db
 
A survey on challenges to the media cloud
A survey on challenges to the media cloudA survey on challenges to the media cloud
A survey on challenges to the media cloud
 
A survey of provenance leveraged
A survey of provenance leveragedA survey of provenance leveraged
A survey of provenance leveraged
 
A survey of private equity investments in kenya
A survey of private equity investments in kenyaA survey of private equity investments in kenya
A survey of private equity investments in kenya
 
A study to measures the financial health of
A study to measures the financial health ofA study to measures the financial health of
A study to measures the financial health of
 

Dernier

Architecting Cloud Native Applications
Architecting Cloud Native ApplicationsArchitecting Cloud Native Applications
Architecting Cloud Native Applications
WSO2
 
Cloud Frontiers: A Deep Dive into Serverless Spatial Data and FME
Cloud Frontiers:  A Deep Dive into Serverless Spatial Data and FMECloud Frontiers:  A Deep Dive into Serverless Spatial Data and FME
Cloud Frontiers: A Deep Dive into Serverless Spatial Data and FME
Safe Software
 

Dernier (20)

Apidays New York 2024 - The value of a flexible API Management solution for O...
Apidays New York 2024 - The value of a flexible API Management solution for O...Apidays New York 2024 - The value of a flexible API Management solution for O...
Apidays New York 2024 - The value of a flexible API Management solution for O...
 
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
 
Navi Mumbai Call Girls 🥰 8617370543 Service Offer VIP Hot Model
Navi Mumbai Call Girls 🥰 8617370543 Service Offer VIP Hot ModelNavi Mumbai Call Girls 🥰 8617370543 Service Offer VIP Hot Model
Navi Mumbai Call Girls 🥰 8617370543 Service Offer VIP Hot Model
 
Apidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, Adobe
Apidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, AdobeApidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, Adobe
Apidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, Adobe
 
Strategize a Smooth Tenant-to-tenant Migration and Copilot Takeoff
Strategize a Smooth Tenant-to-tenant Migration and Copilot TakeoffStrategize a Smooth Tenant-to-tenant Migration and Copilot Takeoff
Strategize a Smooth Tenant-to-tenant Migration and Copilot Takeoff
 
Real Time Object Detection Using Open CV
Real Time Object Detection Using Open CVReal Time Object Detection Using Open CV
Real Time Object Detection Using Open CV
 
Strategies for Landing an Oracle DBA Job as a Fresher
Strategies for Landing an Oracle DBA Job as a FresherStrategies for Landing an Oracle DBA Job as a Fresher
Strategies for Landing an Oracle DBA Job as a Fresher
 
TrustArc Webinar - Stay Ahead of US State Data Privacy Law Developments
TrustArc Webinar - Stay Ahead of US State Data Privacy Law DevelopmentsTrustArc Webinar - Stay Ahead of US State Data Privacy Law Developments
TrustArc Webinar - Stay Ahead of US State Data Privacy Law Developments
 
Boost Fertility New Invention Ups Success Rates.pdf
Boost Fertility New Invention Ups Success Rates.pdfBoost Fertility New Invention Ups Success Rates.pdf
Boost Fertility New Invention Ups Success Rates.pdf
 
Data Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt RobisonData Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt Robison
 
AXA XL - Insurer Innovation Award Americas 2024
AXA XL - Insurer Innovation Award Americas 2024AXA XL - Insurer Innovation Award Americas 2024
AXA XL - Insurer Innovation Award Americas 2024
 
ICT role in 21st century education and its challenges
ICT role in 21st century education and its challengesICT role in 21st century education and its challenges
ICT role in 21st century education and its challenges
 
Architecting Cloud Native Applications
Architecting Cloud Native ApplicationsArchitecting Cloud Native Applications
Architecting Cloud Native Applications
 
Corporate and higher education May webinar.pptx
Corporate and higher education May webinar.pptxCorporate and higher education May webinar.pptx
Corporate and higher education May webinar.pptx
 
Web Form Automation for Bonterra Impact Management (fka Social Solutions Apri...
Web Form Automation for Bonterra Impact Management (fka Social Solutions Apri...Web Form Automation for Bonterra Impact Management (fka Social Solutions Apri...
Web Form Automation for Bonterra Impact Management (fka Social Solutions Apri...
 
How to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerHow to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected Worker
 
Ransomware_Q4_2023. The report. [EN].pdf
Ransomware_Q4_2023. The report. [EN].pdfRansomware_Q4_2023. The report. [EN].pdf
Ransomware_Q4_2023. The report. [EN].pdf
 
DBX First Quarter 2024 Investor Presentation
DBX First Quarter 2024 Investor PresentationDBX First Quarter 2024 Investor Presentation
DBX First Quarter 2024 Investor Presentation
 
Cloud Frontiers: A Deep Dive into Serverless Spatial Data and FME
Cloud Frontiers:  A Deep Dive into Serverless Spatial Data and FMECloud Frontiers:  A Deep Dive into Serverless Spatial Data and FME
Cloud Frontiers: A Deep Dive into Serverless Spatial Data and FME
 
Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024
 

11.development of a writer independent online handwritten character recognition system using modified hybrid

  • 1. Computer Engineering and Intelligent Systems www.iiste.org ISSN 2222-1719 (Paper) ISSN 2222-2863 (Online) Vol 3, No.4, 2012 Development of a Writer-Independent Online Handwritten Character Recognition System Using Modified Hybrid Neural Network Model Fenwa O. D.* Department of Computer Science and Engineering, Ladoke Akintola University of Technology, P.M.B 4000, Ogbomoso, Nigeria. *E-mail of the corresponding author: odfenwa@lautech.edu.ng Omidiora E. O. Department of Computer Science and Engineering, Ladoke Akintola University of Technology, P.M.B 4000, Ogbomoso, Nigeria. E-mail: omidiorasayo@yahoo.co.uk Fakolujo O. A. Department of Computer Science and Engineering, Ladoke Akintola University of Technology, P.M.B 4000, Ogbomoso, Nigeria. E-mail: ola@fakolujo.com Ganiyu R. A. Department of Computer Science and Engineering, Ladoke Akintola University of Technology, P.M.B 4000, Ogbomoso, Nigeria. E-mail: ganiyurafiu@yahoo.com Abstract Recognition of handwritten characters has become a difficult problem because of the high variability and ambiguity in the character shapes written by individuals. Some of the problems encountered by researchers include selection of efficient feature extraction method, long network training time, long recognition time and low recognition accuracy. However, many feature extraction techniques have been proposed in literature to improve overall recognition rate although most of the techniques used only one property of the handwritten character. This research focuses on developing a feature extraction technique that combined three characteristics (stroke information, contour pixels and zoning) of the handwritten character to create a global feature vector. A hybrid feature extraction algorithm was developed to alleviate the problem of poor feature extraction algorithm of online character recognition system. Besides, this research work also focused on alleviating the problem of standard backpropagation algorithm based on ‘error adjustment’. A hybrid of modified Counterpropagation and modified optical backpropagation neural network model was developed to enhance the performance of the proposed character recognition system. Experiments were 89
  • 2. Computer Engineering and Intelligent Systems www.iiste.org ISSN 2222-1719 (Paper) ISSN 2222-2863 (Online) Vol 3, No.4, 2012 performed with 6200 handwriting character samples (English uppercase, lowercase and digits) collected from 50 subjects using G-Pen 450 digitizer and the system was tested with 100 character samples written by people who did not participate in the initial data acquisition process. The performance of the system was evaluated based on different learning rates, different image sizes and different database sizes. The developed system achieved better performance with no recognition failure, 99% recognition rate and an average recognition time of 2 milliseconds. Keywords: Character recognition, Feature extraction, Neural Network, Counterpropagation, Optical Backpropagation, Learning rate. 1. Introduction The use of neural network for handwriting recognition is a field that is attracting a lot of attention. As the computing technology advances, the benefits of using Artificial Neural Network (ANN) for the purpose of handwriting recognition become more obvious. Hence, new ANN approaches geared toward the task of handwriting recognition are constantly being studied. Character recognition is the process of applying pattern-matching methods to character shapes that has been read into a computer to determine which alpha- numeric character, punctuation marks, and symbols the shapes represent. The classes of recognition systems that are usually distinguished are online systems for which handwriting data are captured during the writing process (which makes available the information on the ordering of the strokes) and offline systems for which recognition takes place on a static image captured once the writing process is over (Anoop and Anil, 2004; Liu et al., 2004; Mohamad and Zafar, 2004; Naser et al., 2009; Pradeep et al., 2011). The online methods have been shown to be superior to their offline counterpart in recognizing handwriting characters due the temporal information available with the former (Pradeep et al., 2011). Handwriting recognition system can further be broken down into two categories: writer independent recognition system which recognizes wide range of possible writing styles and a writer dependent recognition system which recognizes writing styles only from specific users (Santosh and Nattee, 2009). Online handwriting recognition today has special interest due to increased usage of the hand held devices. The incorporation of keyboard being difficult in the hand held devices demands for alternatives, and in this respect, online method of giving input with stylus is gaining quite popularity (Gupta et al., 2007). Recognition of handwritten characters with respect to any language is difficult due to variability of writing styles, state of mood of individuals, multiple patterns to represent a single character, cursive representation of character and number of disconnected and multi-stroke characters (Shanthi and Duraiswamy, 2007). Current technology supporting pen-based input devices include: Digital Pen by Logitech, Smart Pad by Pocket PC, Digital Tablets by Wacom and Tablet PC by Compaq (Manuel and Joaquim, 2001). Although these systems with handwriting recognition capability are already widely available in the market, further improvements can be made on the recognition performances for these applications. The challenges posed by the online handwritten character recognition systems are to increase the recognition accuracy and to reduce the recognition time (Rejean and Sargurl, 2000; Gupta et. al., 2007). Various approaches that have been used by many researchers to develop character recognition systems, these include; template matching approach, statistical approach, structural approach, neural networks approach and hybrid approach. Hybrid approach (combination of multiple classifiers) has become a very active area of research recently (Kittler and Roli, 2000; 2001). It has been demonstrated in a number of applications that using more than a single classifier in a recognition task can lead to a significant improvement of the system’s overall performance. Hence, hybrid approach seems to be a promising approach to improve the recognition rate and recognition accuracy of current handwriting recognition systems (Simon and Horst, 2004). However, Selection of a feature extraction method is probably the single most important factor in achieving high recognition performance in character recognition system (Pradeep et. al., 2011). No matter how sophisticated the classifiers and learning algorithms, poor feature extraction will always lead to poor system performance (Marc et. al., 2001). In furtherance, Fenwa et al.(2012) 90
  • 3. Computer Engineering and Intelligent Systems www.iiste.org ISSN 2222-1719 (Paper) ISSN 2222-2863 (Online) Vol 3, No.4, 2012 developed a feature extraction technique for online character recognition system using hybrid of geometrical and statistical features. Thus, through the integration of geometrical and statistical features, insights were gained into new character properties, since these types of features were considered to be complementary. 2. Research Methodology The five stages involved in developing the proposed character recognition system, which include data acquisition, pre-processing, character processing that comprises feature extraction and character digitization, training and classification using hybrid neural network model and testing, are as shown in Figure 2.2. Experiments were performed with 6200 handwriting character samples (English uppercase, lowercase and digits) collected from 50 subjects using G-Pen 450 digitizer and the system was tested with 100 character samples written by people who did not participate in the initial data acquisition process. The performance of the system was evaluated based on different learning rates, different image sizes and different database sizes. 2.1 Data Acquisition The data used in this work were collected using Digitizer tablet (G-Pen 450) shown in Figure 2.3. It has an electric pen with sensing writing board. An interface was developed using C# to acquire data (character information) such as stroke number, stroke pressure, etc from different subjects using the digitizer tablet. 26 Upper case (A-Z), 26 lower case (a-z) English alphabets and 10 digits (0-9) making a total number of 62 characters. 6,200 characters (62 x 2 x 50) were collected from 50 subjects as each individual was requested to write each of the characters 2 times (this is done to allow the network learn various possible variations of a single character and become adaptive in nature). This serves as the training data set which was the input data that was fed into the neural network. 2.2 Data Preprocessing Pre-processing is done prior to the application of feature extraction algorithms. Pre-processing aims to produce clean character images that are easy for the character recognition systems to operate more accurately. Feature extraction stage relies on the output of this process. The pre processing techniques used in this research work is Grid Resizing. 2.3 Resizing Grid From the interface that was provided, there wasn’t any degree of measurement to determine how small/big the input character should be. Hence, the character written is resized to a matrix size of 5 by 7, 10 by 14 and 20 by 28 for all input characters. This is used to get the universe of discourse which is the shortest matrix that fit the entire character skeleton. The universe of discourse is measured to easily get a uniform matrix size in multiple of 5 by 7, 10 by 14 and 20 by 28 respectively. Any character measured smaller than the required size is considerably resized to the multiple of 5 by 7, 10 by 14 and 20 by 28 conversely any character measured larger than the required size will also be resized to the multiple of 5 by 7, 10 by 14 and 20 by 28. This implies that rows and columns of Zero’s are added or subtracted from the resized image matrix to achieve the required multiple of 5 by 7, 10 by 14 and 20 by 28. 2.4 Feature Extraction Development The goal of feature extraction is to extract a set of features, which maximizes the recognition rate with the least amount of elements. Many feature extraction techniques have been proposed to improve overall recognition rate; however, most of the techniques used only one property of the handwritten character. This research focuses on a feature extraction technique that combined three characteristics (stroke information, contour pixels and zoning) of the handwritten character to create a global feature vector. Hence, a hybrid feature extraction algorithm was developed using Geometrical and Statistical features as shown in Figure 91
  • 4. Computer Engineering and Intelligent Systems www.iiste.org ISSN 2222-1719 (Paper) ISSN 2222-2863 (Online) Vol 3, No.4, 2012 2.4. Integration of Geometrical and Statistical features was used to highlight different character properties, since these types of features are considered to be complementary. 2.4.1 The Developed Hybrid (Geom-Statistical) Feature Extraction Algorithm The Hybrid feature extraction model adopted in this work is as shown in Figure 2.4. Stages of development of the proposed hybrid feature extraction algorithm are as follow: Stage 1: Get the stroke information of the input characters from the digitizer (G- pen 450). These include: (i) Pressure used in writing the strokes of the characters (ii) Number (s) of strokes used in writing the characters (iii) Number of junctions and the location in the written characters (iv) The horizontal projection count of the character. Stage 2: Apply Contour tracing algorithm to trace out the contour of the characters: Stage 3: Develop a modified hybrid zoning algorithm and RUN it on the contours of the characters: Two zoning algorithms were proposed by Vanajah and Rajashekararadhya in 2008 for the recognition of four popular Indian scripts (Kannada, Telugu, Tamil and Malayalam) numeral. These were: Image Centroid and Zone-based (ICZ) Distance Metric Feature Extraction System (Vanajah and Rajashekararadhya, 2008a) and Zone Centroid and Zone-based (ZCZ) Distance Metric Feature Extraction System (Rajashekararadhya and Vanajah, 2008b). The Two Algorithms were modified in terms of: (i) Number of zones being used (25 Zones) as shown in Figure 2.1. (ii) Measurement of the distances from both the Image Centroid and Zone Centroid to the pixels present in each zone. (iii) The area of application (uppercase (A-Z), lowercase (a-z) English alphabet and digit (0- 9)). Few zones are adopted in this research but emphasis is laid on how to effectively measure the pixel densities in each zone. However, pixels pass through the zones at varied distances that is why we measured five distances for each zone at an angle of 200 . Hybrid of Modified ICZ and Modified ZCZ based Distance Metric Feature Extraction Algorithm. Input: Pre processed character image Output: Features for classification and recognition Method Begins Step1: Divide the input image into 25 equal zones. Step2: Compute the input image centroid Step3: Compute the distance between the image centroid to each pixel present in the zone measured at angle 200 Step4: Repeat step 3 for the entire pixel present in the zone. Step5: Compute average distance between these points. 92
  • 5. Computer Engineering and Intelligent Systems www.iiste.org ISSN 2222-1719 (Paper) ISSN 2222-2863 (Online) Vol 3, No.4, 2012 Step6: Compute the zone centroid Step7: Compute the distance between the zone centroid to each pixel present in the zone measured at angle 200. Step8: Repeat step 7 for the entire pixel present in the zone Step9: Compute average distance between these points. Step10: Repeat the steps 3-9 sequentially for the entire zones. Step11: Finally, 2xn (50) such features were obtained for classification and recognition. Method Ends Stage 4: Feed the outputs of the extracted features of the characters into the digitization stage in order to convert all the extracted features into digital forms. 2.4.2 Development of Hybrid Neural Network Model In this research work, hybrid approach with serial combination scheme was adopted. Hybrid of modified Counterpropagation and modified Optical Backpropagation neural network was developed as shown in Figure 2.5. The architecture of the neural network with three layers is illustrated in Figure 2.5. The output of ith layer is given by (2.1) except the output layer which uses maxsoft function: ai = logsig(wi ai-1 + bi) (2.1) where i = (2, 3) and a0 = P, a1 = E where E is the Euclidean distance between the weight vector and the input vector. wi = Weight vector of ith layer ai = Output of ith layer bi = Bias vector for ith layer. The input vector ‘P’ is represented by the solid vertical bar at the left. The dimensions of ‘P’ are displayed as 35 x 1, indicating that the input is a single vector of 35 elements (i.e. the image size). These inputs go to weight matrix ‘W1’, which has 86 rows (i.e. 86 neurons in the first hidden layer) and 35 columns. A constant ‘1’ enters the neuron as input and is multiplied by a bias ‘b1’. The net input to the transfer function (Euclidean distance) in the Kohonen (hidden layer) is ‘n1’, which is given as the Euclidean distance between the weight vector wi and the input vector P. The neurons output ‘a1’ serves as inputs to the second hidden layer. These inputs go to weight matrix ‘W2’, which has 86 rows (i.e.86 neurons in the first hidden layer) and 35 columns. A constant ‘1’ enters the neuron as input and is multiplied by a bias ‘b2’. The net input to the transfer function (log sigmoid) in the second (hidden layer) is ‘n2’, which is the sum of the bias ‘b2’ and the product ‘W2a1’. The output ‘a2’ is a single vector of 86 elements and this serves as inputs to the output layer ‘a3’. These inputs go to weight matrix ‘W3’, which has 86 rows (i.e. 86 neurons from the second hidden layer) and 62 columns (26 uppercase + 26 lowercase + 10 digits). The net input to the transfer function (maxsoft) in the output layer is ‘n3’, which is the sum of the bias ‘b3’ and the product ‘W3a2’. The neurons output ‘a3’ is a single vector of 62 elements and this serves as the final output of the neural network. This research work adopted the modified Counterpropagation network (CPN) developed by Jude, Vijila, and Anitha (2010). The training algorithm involves the following two phases: (i) Weight adjustment between the input layer and the hidden layer (Kohonen layer) The weight adjustment procedure for the hidden layer weights is same as that of the conventional CPN. It follows the unsupervised methodology to obtain the stabilized weights. After convergence, the weights between the hidden layer and the output layer are calculated. 93
  • 6. Computer Engineering and Intelligent Systems www.iiste.org ISSN 2222-1719 (Paper) ISSN 2222-2863 (Online) Vol 3, No.4, 2012 (ii) Weight adjustment between the hidden layer and its output layer The weight adjustment procedure employed in this work is significantly different from the conventional CPN. The weights are calculated in the reverse direction without any iterative procedures. Normally, the weights are calculated based on the criteria of minimizing the error. But in this work, a minimum error value is specified initially and the weights are estimated based on the error value. Thus without any training methodology, the weight values are estimated. This technique accounts for higher convergence rate since one set of weights are estimated directly. It is this output that served as the input to the modified optical backpropagation algorithm. 2.4.2.1 Modified Optical Backpropagation Neural Network In the standard backpropagation, the error at a single output unit is defined according to Equation as: δo pk = (Ypk - Opk ) . fo' k (neto pk) (2.2) where the subscript “p” refers to the pth training vector, and “k” refers to the kth output unit, Ypk is the desired output value, Opk is the actual output from kth unit, then δopk will propagate backward to update the output-layer weights and the hidden-layer weights. In the Optical backpropagation (OBP), error at a single output unit is adjusted according to Otair and Salameh (2005) as: New δopk = (1+e (Ypk - Opk )2 . fo'k (neto pk)), if (Y – O) >= zero (2.3a) New δopk = - (1+e (Ypk - Opk )2 . fo'k (neto pk)), if (Y – O) < zero (2.3b) The error function defined in Optical Backpropagation (Otair and Salameh, 2005) earlier is proportional to the square of the Euclidean distance between the desired output and the actual output of the network for a particular input pattern. As an alternative, other error functions whose derivatives exist and can be calculated at the output layer can replace the traditional square error criterion (Haykin, 2003). In this research work, error of the third order (Cubic error) has been adopted to replace the traditional square error criterion used in Optical backpropagation. The Equation of the cubic error is given as: δopk = -3(Ypk – Opk)2 . fo'k(netopk) (2.4) The cubic error in Equation (2.4) was manipulated mathematically in order to maximize the error of each output unit which will be transmitted backward from the output layer to each unit in the intermediate layer. These are as shown in Equations (2.5a) and (2.5b) below: Modified δopk = 3((1+ et)2 . fo'k(netopk)) If (Ypk – Opk)2 >= 0 (2.5a) Modified δopk = -3((1+ et)2 . fo'k(netopk)) If (Ypk – Opk)2 <=0 (2.5b) where Ypk = Target or Desired output Opk = Network output t = (Ypk – Opk)2 However, one of the ways to reduce the training time is through the use of momentum, as it enhances the stability of the training process. The momentum is used to keep the training process going in the same general direction (Haykin, 2003). In the modified Optical Backpropagation network, momentum was introduced. Hence, the weight update for the output unit is: Wokj(t+1) = Wokj(t) + µWokj(t) + (η . Modified δopk. ipj) (2.6) where µ is the momentum coefficient typically about 0.9 and η is the learning rate. 2.4.2.1 The Modified Optical Backpropagation Algorithm Modifications of the algorithm are in terms of: 94
  • 7. Computer Engineering and Intelligent Systems www.iiste.org ISSN 2222-1719 (Paper) ISSN 2222-2863 (Online) Vol 3, No.4, 2012 (i) Error signal function (ii) Application area With the introduction of Cubic error function and Momentum, the modified Optical Backpropagation is given as: 1. Apply the input example to the input units. 2. Calculate the net-input values to the hidden layer units. 3. Calculate the outputs from the hidden layer. 4. Calculate the net-input values to the output layer units 5. Calculate the outputs from the output units 6. Calculate the error term for the output units, using Equations (2.5a) and (2.5b) 7. Calculate the error term for the hidden units, through applying Modified δopk also Modified δh pj = fh'j(neth pj) .( δopk . Wokj) (2.7) 8. Update weights on the output layer. Wokj(t+1) = Wokj(t) + µWokj(t) + (η . Modified δopk . ipj) (2.8) 9. Update weights on the hidden layer. Whji(t+1) = Whji(t) + (η . Modified δhpj . Xi) (2.9) Repeat steps from step 1 to step 9 until the error (Ypk – Opk) is acceptably small for each of the training vector pair. The proposed algorithm as in OBP is stopped when the cubes of the differences between the actual and target values summed over units and all patterns are acceptably small. 2.4.3 The Hybrid Neural Network Algorithm This research work employed hybrid of modified Counterpropagation and modified Optical Backpropagation neural networks for the training and classification of the input pattern. The training algorithm involves the following two stages: Stage A: Performs the training of the weights from the input nodes to the Kohonen hidden node. Step 1: Weight adjustment between the input layer and the hidden layer The weight adjustment procedure for the hidden layer weights is same as that of the conventional CPN. It follows the unsupervised methodology to obtain the stabilized weights. This process is repeated for a suitable number of iterations and the stabilized set of weights are obtained. After convergence, the weights between the Kohonen hidden layer and the output layer are calculated. Step 2: Weight adjustment between the hidden layer and the output layer The weight adjustment procedure employed in this work is significantly different from the conventional CPN. The weights are calculated in the reverse direction without any iterative procedures. Normally, the weights are calculated based on the criteria of minimizing the error. But in this work, a minimum error value is specified initially and the weights are estimated based on the error value. The detailed steps of the modified algorithm are given below. Step 1: The stabilized weight values are obtained when the error value (target output) is equal to zero (or) a predefined minimum value. The error value used for convergence in this work is 0.1. The following procedure uses this concept for weight matrices calculation. Step 2: Supply the target vectors t1 to the output layer neurons Step 3: Since ( t1 – y1 ) = 0.1 for convergence (2.10) 95
  • 8. Computer Engineering and Intelligent Systems www.iiste.org ISSN 2222-1719 (Paper) ISSN 2222-2863 (Online) Vol 3, No.4, 2012 The output of the output layer neurons is set equal to the target values as: y1 = t1 − 0.1 (2.11) Step 4: Once the output value is calculated, the sum of the weighted input signals. Thus without any training methodology, the weight values are estimated. This technique accounts for higher convergence rate since one set of weights are estimated directly. STAGE B: Performs the training of the weights from the second hidden node to the output nodes. 1. Calculate the net-input values from the Kohonen to the second hidden layer units. 2. Calculate the outputs from the second hidden layer. 3. Calculate the net-input values to the output layer units 4. Calculate the outputs from the output units 5. Calculate the error term for the output units, using Equations (2.5a) and (2.5b) 6. Calculate the error term for the hidden units, through applying modified δopk as in Equation (2.7) 7. Update weights on the output layer using (2.8) 8. Update weights on the hidden layer using Equation (2.9) Repeat steps from step 1 to step 8 until the error (Ypk – Opk) is acceptably small for each of the training vector pair. The proposed algorithm as classical BP is stopped when the cubes of the differences between the actual and target values summed over units and all patterns are acceptably small. The description of notation used in the training procedure is as given below: Xpi : net input to the ith input unit nethpj: net input to the jth hidden unit Whji: weight on the connection from the ith input unit to jth hidden unit ipj: net input to the jth hidden unit netopk: net input to the kth output unit Wokj: weight on the connection from the jth hidden unit to kth output unit Opk: actual output for the kth output unit 3. Software Implementation This section discussed the phases of developing the proposed character recognition system. All the algorithms have been implemented using C# programming language and RUN under Windows7 operating system on Pentium (R) 2.00GB RAM, 1.83GH Processor and Pentium (R) 4.00GB RAM, 2.13GH Processor respectively. Different interfaces representing the phases of the system development are shown from Figures 2.6 to 2.11. 3.1 Character Acquisition Character acquisition interface as in Figures 2.7 and 2.8 displayed a drawing area to serve as a platform to acquire characters from users. Any drawn character on this platform must be saved into the developed system’s database by clicking the ‘Add Botton’. It also captured number of strokes and pressure used in writing a character. A total 6,200 character samples were acquired using G-Pen 450 as shown in Figure 2.2 96
  • 9. Computer Engineering and Intelligent Systems www.iiste.org ISSN 2222-1719 (Paper) ISSN 2222-2863 (Online) Vol 3, No.4, 2012 and stored in the database. 3.2 Network Training Before the network can be trained, the network parameters such as the learning rate, epoch value, the quit error and character image size must be specified in the ‘Application setting’ interface. The next thing is to specify the total number of the character to be loaded from the database for training. The database size can be varied by specifying the maximum number of character to be selected from character category (uppercase, lowercase and digit). This can be accomplished with the interfaces shown in Figure 2.9 and Figure 2.10. On clicking the ‘Training phase button’, the neural network trains itself with alphanumeric characters of A – Z, a – z and 0 – 9 by performing several iterations based on the error value, learning rate and the epoch value stipulated for the Neural Network. Epoch increases from 1 to 10,000 and the Error reduces from a particular value depending on the value specified by the user. The training is terminated when either the epoch reaches 10,000 or the error reaches the minimum value specified. Network training results such the training time, the epoch value were displayed as shown in Figure 2.10. 3.3 Recognition Phase Recognition of the character can be accomplished by clicking the ‘Data acquisition/ Recognition phase botton’. User is required to write a character and click the ‘Recognize botton’. The results such as the ‘detected character’, ‘recognition rate’ and ‘recognition time’ will be displayed as shown in Figure 2.11. 4. Performance Evaluation The performance metrics adopted in evaluating the developed system are: different learning rate parameters, different image sizes, different database sizes and system configurations. The results are as given in Figure 2.12 to Figure 2.17. Figure 2.12 shows the graph of learning rate versus epoch. Learning rate parameter variation has a positive effect on the network performance. The smaller the value of learning rate, the lower the value with which the network updates its weights. This intuitively implies that it will be less likely to face the over learning difficulty since it will be updating its links slowly and in a more refined manner. However, this would also imply more number of iterations is required to reach its optimal state. Figure 2.13 indicates the graph of image size versus epoch. The image size is directly proportional to epoch. Usually, the complex and large sized input sets require a large topology network with more number of iterations and this has an adverse effect on the recognition time. Figure 2.14 shows the graph of variation of database size and recognition rates. Increase in database size has a positive proportionality relation to the recognition rates due to the fact that network is able to attribute the test character to larger character samples in the vector space. However, rate of increment in recognition rate with respect to database size is considerably small. Moreover, it can be shown from Figure 2.15 that the more the dimensional input vector (image matrix size), the less the recognition performance. Three different image sizes (5 by 7, 10 by 14 and 20 by 28) were considered; it was shown from the results that the higher the image size, the higher the percentage of recognition, although, the rate of change was small. Furthermore, increase in image size also increases the recognition time. System configuration is also another factor influencing the performance of the recognition system. Both the training time and recognition time are measured in CPU (processor) seconds, the higher the speed of the system (processor speed), the lower the training time and the recognition time. This is due to the fact that more time will be used by 1.8GH system to reach the required epochs (iteration) than the system with 2.13GH speed. The result is as shown in Figure 2.16. Figure 2.17 is the graph showing comparison of performance of the developed system with related works in literature. It can be shown from the graph that the accuracy of the system developed by Muhammad et. al. (2005) is higher than that of Muhammad et. al. (2006). This shows that the performance of Counterpropagation neural network is better than that of Standard Backpropagation. The best recognition rate was achieved in the developed system with no 97
  • 10. Computer Engineering and Intelligent Systems www.iiste.org ISSN 2222-1719 (Paper) ISSN 2222-2863 (Online) Vol 3, No.4, 2012 recognition failure. This means that the developed system was able to recognise characters irrespective of the writing styles. 5. Conclusion and Future Work In this paper, we have been able to develop an effective feature extraction technique for the proposed online character recognition system using hybrid of geometrical and statistical features. The hybrid feature extraction was developed to alleviate the problem of poor feature extraction algorithm of online character recognition system. However, a hybrid of modified counter propagation and modified optical backpropagation neural network model was developed for the proposed system. The performance of the online character recognition system under consideration was evaluated based on different learning rates, image sizes and database sizes. The results from the study were compared with some works in the literature using counterpropagation and standard backpropagation respectively. The results showed that counterpropagation performed better than standard backpropagation in terms of correct recognition, false recognition, recognition failure and the achieved recognition rates were 94% and 81% respectively. The developed system achieved better performance with no recognition failure, 99% recognition rate and an average recognition time of 2 milliseconds. Future work could be geared towards integrating an optimization algorithm into the learning algorithms to further enhance the convergence of the neural network. References Anoop, M. N. and Anil K.J. (2004): “Online Handwritten Script Recognition’’, IEEE Trans. PAMI, 26(1): 124 130. Liu, C.L., Nakashima, K., Sako, H. and Fujisawa, H. (2004): “Handwritten Digit Recognition: Investigation of Normalization and Feature Extraction Techniques’’, Pattern Recognition, 37(2): 265-279. Mohamad, D. and Zafar, M.F. (2004): “Comparative Study of Two Novel Feature Vectors for Complex Image Matching Using Counterpropagation Neural Network’’, Journal of Information Technology, FSKSM, UTM, 16(1): 2073-2081. Naser, M.A., Adnan, M., Arefin, T.M., Golam, S.M. and Naushad, A. (2009): “Comparative Analysis of Radon and Fan-beam based Feature Extraction Techniques for Bangla Character Recognition”, IJCSNS International Journal of Computer Science and Network Security, 9(9): 120-135. Pradeep, J., Srinivasan, E. and Himavathi, S. (2011): “Diagonal Based Feature Extraction for Handwritten Alphabets Recognition using Neural Network’’, International Journal of Computer Science and Information Technology (IJCS11), 3(1): 27-37. Shanthi, N., and Duraiwamy, K. (2007): “Performance Comparison of Different Image Sizes for Recognizing Unconstrained Handwritten Tamil Characters Using SVM’’, Journal of Science, 3(9): 760-764. Santosh, K.C. and Nattee, C. (2009): “A Comprehensive Survey on Online Handwriting Recognition Technology and Its Real Application to the Nepalese Natural Handwriting’’, Kathmandu University Journal of Science, Engineering Technology, 5(1): 31-55. Simon G. and Horst B. (2004): “Feature Selection Algorithms for the Generalization of Multiple Classifier Systems and their Application to Handwritten Word Recognition’’, Pattern Recognition Letters, 25(11): 1323-1336. 98
  • 11. Computer Engineering and Intelligent Systems www.iiste.org ISSN 2222-1719 (Paper) ISSN 2222-2863 (Online) Vol 3, No.4, 2012 Gupta, K., Rao, S.V., and Viswanath (2007): “Speeding up Online Character Recognition’’, Proceedings of Image and Vision Computing Newzealand, Hamilton: 41-45. Manuel, J., Fonseca, and Joaquim, A.J., (2001): “Experimental Evolution of an Online Scribble Recognizer’’, Pattern Recognition Letters, 22(12): 1311-1319. Rejean, P. and Sargur, S.N. (2000): “On-line and Off-line Recognition: A Comprehensive Survey’’, IEEE Transaction on Pattern Analysis and Machine Intelligence, 22(1): 63-84. Kittler, J. and Roli, F. (2000): “1st International Workshop on Multiple Classifier Systems”, Cagliari, Italy. Kittler, J. and Roli, F. (2001): “2nd International Workshop on Multiple Classifier Systems”, Cagliari, Italy. Marc, P., Alexandre, L., and Christian, G. (2001): “Character Recognition Experiment using Unipen Data’’, Parizeau and al., Proceeding of ICDAR, Seattle: 10-13. Muhammad, F.Z., Dzuulkifli, M., and Razib, M.O. (2006): “Writer Independent Online Handwritten Character Recognition Using Simple Approach’’, Information Technology Journal, 5(3): 476-484. Freeman, J.A. and Skapura, D.M. (1992): “Backpropagation Neural Networks Algorithm Applications and Programming Techniques”, Addison-Wesley Publishing Company: 89-125. Minai, A.A. and Williams, R.D. (1990): “Acceleration of Backpropagation through learning Rate Momentum Adaptation’’, Proceeding of the International Joint Conference on Neural Networks: 1676-1679. Riedmiller, M. and Braun, H. (1993): “A Direct Adaptive Method for Faster Backpropagation learning the PROP Algorithm’’, Proceedings of the IEEE International Conference on Neural Networks (ICNN), 1: 586- 591, Francisco. Otair, M.A. and Salameh, W.A. (2005a): “Online Handwritten Character Recognition using an Optical Backpropagation Neural Network’’, Issues in Informing Science and Information Technology, 2: 787-797. Rajashekararadhya, S.V. and Vanaja, P.R. (2008a): “Handwritten numeral recognition of three popular South Indian Scripts: A Novel Approach:’’, Proceedings of the Second International Conference on information processing ICIP: 162-167. Rajashekararadhya S.V. and Vanaja, P.R (2008b): “Isolated Handwritten Kannada Digit Recognition: A Novel Approach’’, Proceedings of the International Conference on Cognition and Recognition: 134-140. Fenwa, O. D., Omidiora, E. O. and Fakolujo O. A. (2012): “Development of a Feature Extraction Technique for Online Character Recognition System, Journal of Innovative Systems Design and Engineering, International Institute of Science, Technology and Education,Vol. 3, No.3, pp. 10-23. 99
  • 12. Computer Engineering and Intelligent Systems www.iiste.org ISSN 2222-1719 (Paper) ISSN 2222-2863 (Online) Vol 3, No.4, 2012 Figure 2.1: Character ‘n’ in 5 by 5 (25 equal zones) Figure 2.2: Block Diagram of the Developed Character Recognition System Figure 2.3: The snapshot of Genius Pen (G-Pen 450) Digitizer for character acquisition 100
  • 13. Computer Engineering and Intelligent Systems www.iiste.org ISSN 2222-1719 (Paper) ISSN 2222-2863 (Online) Vol 3, No.4, 2012 Figure 2.4: The Developed Hybrid Feature Extraction Model a1 = E (W1 – P), a2 = logsig (W2a1 + b2), a3 = Maxsoft (W3 + b3) Figure 2.5: The Hybrid Neural Network Model 101
  • 14. Computer Engineering and Intelligent Systems www.iiste.org ISSN 2222-1719 (Paper) ISSN 2222-2863 (Online) Vol 3, No.4, 2012 Figure 2.6: Graphic user interface of the developed Online Character Recognition System Figure 2.7: Acquisition of character ’A’ using 3 strokes 102
  • 15. Computer Engineering and Intelligent Systems www.iiste.org ISSN 2222-1719 (Paper) ISSN 2222-2863 (Online) Vol 3, No.4, 2012 Figure 2.8: Acquisition of character ’A’ using 2 strokes Figure 2.9: Network parameter setting and loading data to be trained from the database Figure 2.10: Network Training in progress Figure 2.11: Result of recognition process displaying recognition status 103
  • 16. Computer Engineering and Intelligent Systems www.iiste.org ISSN 2222-1719 (Paper) ISSN 2222-2863 (Online) Vol 3, No.4, 2012 Figure 2.12: Graph showing the effect of variation in Learning Rate and Database size on Epochs Figure 2.13: Graph showing the effect of variation of Image size and Database size on Epoch Figure 2.14: Graph showing the effect of variation in Database size and Epoch on the Recognition Rate 104
  • 17. Computer Engineering and Intelligent Systems www.iiste.org ISSN 2222-1719 (Paper) ISSN 2222-2863 (Online) Vol 3, No.4, 2012 Figure 2.15: Graph showing the effect of variation in Image size and Database size on the Recognition Rate Figure 2.16: Graph showing the effect of variation in system configurations and Database size on Epochs CR = Correct Recognition; FR = False Recognition; RF = Recognition Failure Figure 2.17: Graph showing performance evaluation of the developed system with related works 105
  • 18. International Journals Call for Paper The IISTE, a U.S. publisher, is currently hosting the academic journals listed below. The peer review process of the following journals usually takes LESS THAN 14 business days and IISTE usually publishes a qualified article within 30 days. Authors should send their full paper to the following email address. More information can be found in the IISTE website : www.iiste.org Business, Economics, Finance and Management PAPER SUBMISSION EMAIL European Journal of Business and Management EJBM@iiste.org Research Journal of Finance and Accounting RJFA@iiste.org Journal of Economics and Sustainable Development JESD@iiste.org Information and Knowledge Management IKM@iiste.org Developing Country Studies DCS@iiste.org Industrial Engineering Letters IEL@iiste.org Physical Sciences, Mathematics and Chemistry PAPER SUBMISSION EMAIL Journal of Natural Sciences Research JNSR@iiste.org Chemistry and Materials Research CMR@iiste.org Mathematical Theory and Modeling MTM@iiste.org Advances in Physics Theories and Applications APTA@iiste.org Chemical and Process Engineering Research CPER@iiste.org Engineering, Technology and Systems PAPER SUBMISSION EMAIL Computer Engineering and Intelligent Systems CEIS@iiste.org Innovative Systems Design and Engineering ISDE@iiste.org Journal of Energy Technologies and Policy JETP@iiste.org Information and Knowledge Management IKM@iiste.org Control Theory and Informatics CTI@iiste.org Journal of Information Engineering and Applications JIEA@iiste.org Industrial Engineering Letters IEL@iiste.org Network and Complex Systems NCS@iiste.org Environment, Civil, Materials Sciences PAPER SUBMISSION EMAIL Journal of Environment and Earth Science JEES@iiste.org Civil and Environmental Research CER@iiste.org Journal of Natural Sciences Research JNSR@iiste.org Civil and Environmental Research CER@iiste.org Life Science, Food and Medical Sciences PAPER SUBMISSION EMAIL Journal of Natural Sciences Research JNSR@iiste.org Journal of Biology, Agriculture and Healthcare JBAH@iiste.org Food Science and Quality Management FSQM@iiste.org Chemistry and Materials Research CMR@iiste.org Education, and other Social Sciences PAPER SUBMISSION EMAIL Journal of Education and Practice JEP@iiste.org Journal of Law, Policy and Globalization JLPG@iiste.org Global knowledge sharing: New Media and Mass Communication NMMC@iiste.org EBSCO, Index Copernicus, Ulrich's Journal of Energy Technologies and Policy JETP@iiste.org Periodicals Directory, JournalTOCS, PKP Historical Research Letter HRL@iiste.org Open Archives Harvester, Bielefeld Academic Search Engine, Elektronische Public Policy and Administration Research PPAR@iiste.org Zeitschriftenbibliothek EZB, Open J-Gate, International Affairs and Global Strategy IAGS@iiste.org OCLC WorldCat, Universe Digtial Library , Research on Humanities and Social Sciences RHSS@iiste.org NewJour, Google Scholar. Developing Country Studies DCS@iiste.org IISTE is member of CrossRef. All journals Arts and Design Studies ADS@iiste.org have high IC Impact Factor Values (ICV).