SlideShare une entreprise Scribd logo
1  sur  8
Télécharger pour lire hors ligne
International Journal of Electrical and Computer Engineering (IJECE)
Vol. 10, No. 1, February 2020, pp. 778~785
ISSN: 2088-8708, DOI: 10.11591/ijece.v10i1.pp778-785  778
Journal homepage: http://ijece.iaescore.com/index.php/IJECE
Application on character recognition system on road sign for
visually impaired: case study approach and future
Jaejoon Kim
School of Computer and Communication, Daegu University, South Korea
Article Info ABSTRACT
Article history:
Received Apr 7, 2019
Revised Sep 18, 2019
Accepted Sep 27, 2019
Many visually impaired people worldwide are unable to travel safely and
autonomously because they are physically unable to perceive effective visual
information during their daily lives. In this research, we study how to extract
the character information of the road sign and transmit it to the visually
impaired effectively, so they can understand easier. Experimental method is
to apply the Maximally Stable External Region and Stroke Width Transform
method in Phase I so that the visually impaired person can recognize
the letters on the road signs. It is to convey text information to the disabled.
The result of Phase I using samples of simple road signs was to extract
the sign information after dividing the exact character area, but the accuracy
was not good for the Hangul (Korean characters) information. The initial
experimental results in the Phase II succeeded in transmitting the text
information on Phase I to the visually impaired. In the future, it will be
required to develop a wearable character recognition system that can be
attached to the visually impaired. In order to perform this task, we need to
develop and verify a miniaturized and wearable character recognition system.
In this paper, we examined the method of recognizing road sign characters on
the road and presented a possibility that may be applicable to our final
development.
Keywords:
Character recognition system
MSER
OCR
Visually impaired
Wearable device
Copyright © 2020 Institute of Advanced Engineering and Science.
All rights reserved.
Corresponding Author:
Jaejoon Kim
School of Computer and Communication,
Daegu University,
201 Daegudae-ro, Gyeongsan, Gyeongbuk, 38453, South Korea.
Email: jjkimisu@daegu.ac.kr
1. INTRODUCTION
There are many people with visual disabilities in the world as well as people reside in Korea.
Auxiliary devices have been actively developed to provide visually impaired persons with accurate
information on road signs during walking, which can be proven as crucial parts of everyday social activities.
Even more, various devices such as products that read or provide directions for blind have been actively
developed in the past.
Various methods are currently in development to automatically extract the region surrounding
a character when that character is detected to be in the presence of a visually impaired person [1-5].
The method consists of mainly first recognizing the character string in its region and then sending
the scanned information back to the user. Due to artificial intelligence technology, additional features can
include a dictionary function in case the user does not understand the meaning of the sign in question.
In recent years, various techniques using artificial intelligence in image recognition, natural language
processing, and natural language generation have been developed to help people with visual impairments to
maintain a comprehensive and productive life [6].
Int J Elec & Comp Eng ISSN: 2088-8708 
Application on character recognition system on road sign for visually impaired: case ... (Jaejoon Kim)
779
In this paper, we tried to develop a variety of wearable assistive aids for the visually impaired to
eliminate the inconvenience of walking or real life. In order to do this, in Phase I, Maximally Stable External
Region (MSER) and Stroke Width Transform (SWT) features were applied to character recognition of road
signs and pedestrian signs [7, 8]. In Phase II, we tried to design a text recognition system based on TTS
(Text-to-Speech) converter [9] that transmits the text recognition information to the visually impaired.
This paper proposes a plan for the visually impaired to meet these developments and to present
the applicable possibilities. Section 2 briefly discusses the related works for character string recognition
technology currently in development for the visually impaired and characteristics of information guiding
system. Section 3 describes experimental implementation of character recognition for our proposed system.
Sections 4 and 5 conclude the paper by discussing experimental results, future works and other possible
solutions.
2. RELATED WORK
2.1. Characteristics and trend of information guiding system
Blind people are divided into low vision and the blind. Low vision refers to vision problems that
cannot be improved by medical or optical means due to birth defects or acquired eye diseases. According to
a report of Yonsei University Medical Center, the suicide rate of the low vision is twice that of the normal
person, and the psychological stress due to the low vision is large, and the need for an auxiliary device to
help the low vision is increasing [10]. In addition, a report from the UN estimates that the world‟s population
age 60 or older will increase from 11% in 2012 to 22% by 2020 as shown in Figure 1. Especially in North
America and Europe, the distribution of population aged 60 years or older is reported to be significantly
higher than the average [11]. In the point of world market, Figure 2 indicated that the global elderly and
disabled assistive devices market was valued at US$ 14,109.1 million in 2015 and is expected to expand at
a CAGR of 7.4% during the forecast period (2016 – 2024) [12].
Figure 1. The statistcs of UN, gendered innovations (2012 vs. 2050) [11]
Figure 2. Global elderly and disabled assistive devices market size and forecast (unit: US$ Million) [12]
 ISSN: 2088-8708
Int J Elec & Comp Eng, Vol. 10, No. 1, February 2020 : 778 - 785
780
2.2. Character information recognition system
To allow a visually impared person to recognize a character, our application must be able to find
the character region in which the sign or the message is written. The most well-known method used for
binarization is a global binarization which is based on a single threshold [13-15]. However, these methods
have chracteristics such as performance issues with the overall brightness of the input document image are
not constant.
Another method used for binarization is the Maximally Stable External Region (MSER) algorithm,
which is used in a more robust manner of finding a region with a small rate of change while changing the
threshold value in order to distinguish characters in both bright and dark regions. The character area is
generally easy to distinguish from the background and has a constant brightness value. This characteristic of
the character region is suitable to be detected by the MSER method [7, 8]. Various applications have been
developed by various companies, all of which attempt to extract a character string by extracting horizontal
and vertical boundary components from the input image. These methods basically either find the portion with
the greatest boundary component or extract a character string that is irrespective of misjudgment or
distortion.
The ABBYY FineReader Engine, OCR SDK (character recognition development toolkit) supplied
by Retia, was applied to the development of a 'print-to-speech converter', which is a device for recognizing
printed characters for visually impaired and converting them into voice [16-18]. ABBYY [19] developed
ABBYY FineReader Engine with a goal of a stable „print-to-speech‟ converter for the visually impaired.
The OCR key technologies of ABBYY FineReader Engine have several main features such as:
a. Korean, Chinese, English, numbers and special characters support
b. Vertical reading, Multi reading
c. Batch scanning, batch conversion possible
d. It can save word, text file etc.
By utilizing this engine, the visually impaired user can convert the desired document to speech
through this device. While in the past books, receipts and other readable documents could not be read without
Braille documents, the engine now makes it possible to directly scan a desired document or image by using
the „sound‟ with OCR technology and convert it so the user can hear the by voice. The „finger reader‟ [20]
product developed by MIT works by placing a finger near the book. A high-resolution camera is used to scan
the characters corresponding to the finger‟s touch while also reading them out loud after being scanned. After
several years of research, OrCam, an Israeli company, is now developing new products that use facial
recognition technology to scan not only characters but also various colors [21]. There are also „smart glasses‟
for people with partial vision loss. These „smart glasses‟ work by viewing images in 3D glasses. However
even though these glasses have been proven useful reducing inconvenience of the visually disabled, they are
still very expensive. Therefore there will be difficulties in making them commercially available for everyone.
Figure 3 shows the number of people in Korea with sight disabilities is around about 250,000 people as of
2016. Unfortunately, the number of people with visual disabilities is growing larger each year by more than
10%. In addition, the types of disabilities are discribed in Figure 4.
Figure 3. The statistics of disabilities in Korea, 2007 and 2016 [22]
Int J Elec & Comp Eng ISSN: 2088-8708 
Application on character recognition system on road sign for visually impaired: case ... (Jaejoon Kim)
781
Figure 4. The types of visually impaired. (a) Scenery seen with normal eye, (b) Scenery seen with eyes with
cataracts, (c) Scenery seen with eyes with glaucoma, (d) Scenery seen with eyes with diabetic retinopathy,
(e) Scenery seen with eyes with macular degeneration [23]
3. EXPERIMENTAL IMPLEMENTATION
One of the most important parts of the application is to provide information that is only necessary in
order to never confuse the user. Therefore, the application must be able to select only the relevant area
amongst the other „background noise‟. For example, the application must be able to distinguish between
a road sign and the other natural background surrounding the sign. In this paper, we aim to detect the text
area in the image provided with the visually impaired and to recognize the information provided in the area
as a character and deliver it to the visually impaired. In order to implement our goal, we used MSER and
OCR methods for detecting information area and text detection as a starting point of research. Based on
the Phase I of the research, we aim to develop a wearable device that provides road information about
the visually impaired by utilizing the TTS (Text-to-Speech) function of a typical smartphone.
3.1. Feature extraction with MSER
Information on the road signs do not contain as much textual information when compared to books
or newspapers. As a method of the visually impaired to arrive at their desired destination, they had to
previous rely the usage of a guidance stick. While it may to practical and has been used as from
the beginning of blind people, a different method can be developed. These days it is possible to detect
the text in an image using no more than a smartphone with a camera. By using pixel spans as nodes, blobs
can be generated which can detect text in an image more quickly and accurately. In addition, as the usage of
smartphone increased very rapidly, an optical character recognition application that recognizes a character of
an image captured by a camera mounted on a smart phone and displays it on a display can be distributed in
the near future and prove more practical compared to a guidance stick.
The most important step is to detect the blob which does a candidate region exists in the image that
is most likely to contain the text of the road sign. The SWT algorithm [24, 25] is a method of detecting text
by determining which region has little change in the shape and thickness. When the parameters are given,
the algorithm detects the text within the image using those parameters. The main advantage of SWT are that
it can detect text in an image without a separate learning process, but it also has a drawback in which it takes
a lot of time to detect text in an image because of its complicated operation. The MSER algorithm for robust
text detection is widely used as an algorithm for detecting blob as an aggregate region of pixels different
from surrounding pixels in intensity. The MSER algorithm has the advantage that it can detect the blob faster
than the SWT algorithm, but it has a disadvantage that the accuracy is somewhat lacking, such as detecting
the tickle or noise by blob.
3.2. Application procedure
Figure 5 is an overall block diagram that shows the process of our application after capturing
the image and extracting the text. In this paper, we aimed to test the design possibility of an assistive device
which can be helpful for the visually impaired through Phase I process in two stages. The first step captures
images of a smartphone or future terminal device that the blind has and then removes non-text regions using
MSER. The MSER extraction implements the following steps:
 ISSN: 2088-8708
Int J Elec & Comp Eng, Vol. 10, No. 1, February 2020 : 778 - 785
782
Figure 5. The overall block diagram
a. Sweep threshold of intensity
b. Extract connected components
c. Find a threshold on maximally stable
d. Approximate a region with an ellipse
e. Keep region descriptors
One of the most important features of the MSER algorithm is that it detects most text in an image
but also extracts a stable region, not a text region. To make the algorithm suit our needs, we removed non-
text areas using parameters that distinguish words by their geometrical properties within the image.
In addition, SWT can be used to distinguish more sophisticated non-text areas. In this paper, we applied this
method to find and remove non-text areas. In order to merge individual text areas to extract information on
a single word or a text line after finally confirming the text area, a neighboring text area is searched to form
a bounding box of these areas, and finally, the information is extracted. In Phase II, the final detected text
information is converted into voice information on the visually impaired to be transmitted or applied to future
applicable and wearable devices.
4. RESULT AND DISCUSSION
4.1. Phase I
In our experiment, we used a typical image of the road sign which the visually impaired can easily
encounter on the road. The left column of Figure 6 shows test data set. These images were sorted in order to
make test conditions more difficult. When comparing the amount of text to the whole image there is not
much text contained in the image in the first place. Therefore, these images were ideal in testing if our
proposed method of combining MSER and SWT to extract the necessary information. These test conditions
were also important because the end goal of our application is to distribute it so that it can be used during
everyday activities. Thus, we made sure that the images were selected very carefully in order to simulate
what the user might encounter while using the application.
In Figure 6, we show the results of applying MSER, geometric characteristics, and then applying
SWT to find non-text areas in captured images using mobile phones of the visually impaired. These results
are shown on right column of Figure 6 and Table 1 shows the final detected character information of each
test image. In our test case of images containing only English characters, accuracy was high in
the implemented method. However, when we applied our method to an image where Korean and English
were mixed, we could see a drop off in performance. Through this result we can confirm that a different
detection engine is required for the detection of multiple languages.
Int J Elec & Comp Eng ISSN: 2088-8708 
Application on character recognition system on road sign for visually impaired: case ... (Jaejoon Kim)
783
(a) (b) (c) (d)
(e) (f) (g) (h)
Figure 6. The test data set and experimental results
Table 1. Implemental and detected results
Test data Detected Text Information
Figure 6-(a)
Cyclists
AHEAD
SLOW
Figure 6-(c) AHEAD
Figure 6-(e)
PED
XING
Figure 6-(g)
‟.„i".Si6
SLOW
4.2. Phase II
TTS is a type of speech synthesis program that makes the contents of computer documents, such as
help files and web pages, sound human-readable. TTS can also read image information for people with visual
impairments. There are many TTS products on the market, including Read Please 2000, Proverbe Speech
Unit, and TextAloud. Lucent and AT & T have their own products called “Text-to-Speech.” In this research,
we applied Microsoft Speech API's speech to text functionality in consideration of efficient aspects [26].
Figure 7 shows simplified block diagram for TTS function. This system works well for our test data.
(a) (b) (c)
Figure 7. The block diagram for text-to-speech process. (a) input image,
(b) captured letters on notepad, and (c) speech out with text.txt file
 ISSN: 2088-8708
Int J Elec & Comp Eng, Vol. 10, No. 1, February 2020 : 778 - 785
784
5. CONCLUSION
In this paper, we proposed a character recognition system for visual impaired persons and added in
a voice guidance application implementation process. Our goal is to add a character recognition system for
the visually impaired to the smartphone terminal through H/W production. We proposed a method to detect
the characters and apply the voice guidance from the images captured in Phase I. Using the features of MSER
and SWT, it showed the possibility of extracting the letters of road signs, and in Phase II, it showed the result
of delivering the extracted text information to the users.
However, in this paper, we can look at various future points to consider. So far, the road guidance
service application for the visually impaired has been developed using a blue pitcher communication based
H/W auxiliary device such as a beacon in a specific area. With the development of IoT technology,
the objects and data related to IoT have been increased greatly. It is necessary to develop a camera-based
video information guidance service app that utilizes a voice guidance service technology that provides
appropriate information. In addition, artificial intelligence techniques such as image recognition, natural
language processing, and natural language generation can be used to enable blind people to live a more
comprehensive and productive life.
ACKNOWLEDGEMENTS
This research was supported by the Daegu University Research Grant, 2017.
REFERENCES
[1] A. A. Panchal, et al., “Character detection and recognition system for visually impaired people,” Proceedings of
IEEE International Conference on Recent Trends in Electronics, Information & Communication Technology, India,
pp. 1492-1496, 2016.
[2] Abdurrasyid, et al., “Detection of immovable objects on visually impaired people walking aids,” TELKOMNIKA
Telecommunication Computing Electronics and Control, vol. 17, pp. 580-585, 2019.
[3] D. A. Mohammed, et al., “Off-line handwritten character recognition using an integrated DBSCAN-ANN scheme,”
Indonesian Journal of Electrical Engineering and Computer Science, vol. 14, pp. 1443-1451, 2019.
[4] O. O. Oladayo, “Yoruba Language and Numerals‟ Offline Interpreter Using Morphological and Template
Matching,” IAES International Journal of Artificial Intelligence, vol. 3, pp. 64-72, 2014.
[5] I. F. Bt. Hairuman and O. M. Foong, “OCR signage recognition with skew & slant correction for visually impaired
people,” Proceedings of IEEE International Conference on Hybrid Intelligent Systems, Melacca, Malaysia, pp. 306-
310, 2011.
[6] M. E. Pollack, “Intelligent Technology for an Aging Population: The Use of AI to Assist Elders with Cognitive
Impairment,” AI Magazine, American Association for Artificial Intelligence, vol. 26, pp. 9-21, 2005.
[7] J. Matas, et al., “Robust Wide Baseline Stereo from Maximally Stable Extremal Regions,” Proceedings of the
British Machine Vision Conference, BMVC, pp. 384-393, 2002.
[8] Forssen and Per-Erik, “Maximally Stable Colour Regions for Recognition and Matching,” Proceedings of IEEE
Conference on Computer Vision and Pattern Recognition, CVPR, Minneapolis, USA, 2007.
[9] Q. Ye and D. Doermann, “Text Detection and Recognition in Imagery: A Survey,” IEEE Transactions on Pattern
Analysis and Machine Intelligence, vol. 37, pp. 1480-1500, 2015.
[10] T. H. Rim, et al., “Influence of visual acuity on suicidal ideation, suicide attempts and depression in South Korea,”
British Journal of Ophthalmology, vol. 99, pp. 1112-1119, 2015.
[11] Exploring Markets for Assistive Technologies for the Elderly, UN Gendered Innovations. Available:
https://genderedinnovations.stanford.edu/case-studies/robots.html.
[12] Global Elderly and Disabled Assistive Devices Market, “Coherent Market Insights,” 2017.
[13] N. Otsu, “A Threshold Selection Method from Gray Level Histogram,” IEEE Transactions on System, vol. 19,
pp. 62-66, 1979.
[14] G. Park, et al., “A Study on Enhanced Binarization Method by Using Intensity Information,” Proceedings of
the Spring Conference of the Korea Multimedia Society, pp. 441-445, 2003.
[15] K. Munadi, et al., “Improved Thresholding Method for Enhancing Jawi Binarization Performance,” Proceedings of
14th IAPR International Conference on Document Analysis and Recognition, Kyoto, Japan, vol. 1, pp. 1108-1113,
2017.
[16] R. Smith, “An Overview of the Tesseract OCR Engine,” Proceedings of 9th International Conference on Document
Analysis and Recognition, Brazil, 2007.
[17] A. Abdulkader and M. R. Cassey, “Low cost correction of OCR errors using learning in a multi-engine
environment,” Proceedings of IEEE International Conference on Document Analysis and Recognition, ICDAR2009,
pp. 576-580, 2009.
[18] T. M. Breuel, et al., “High-performance OCR for printed English and fraktur using LSTM networks,” Proceedings
of IEEE International Conference on Document Analysis and Recognition, ICDAR2013. Washington, USA, 2013.
[19] M. Heliński, M. Kmieciak, T Parkoła, “Report on the comparison of Tesseract and ABBYY FineReader OCR
engines,” Impact, 2012.
Int J Elec & Comp Eng ISSN: 2088-8708 
Application on character recognition system on road sign for visually impaired: case ... (Jaejoon Kim)
785
[20] R. Shilkrot, J. Huber, W. M. Ee, P. Maes, S. C. Nanayakkara, “FingerReader: A Wearable Device to Explore
Printed Text on the Go,” Proceedings of the 33rd Annual ACM Conference on Human Factors in Computing
Systems, CHI 2015, Seoul, Korea, 2015.
[21] M. Waisbourd, O. M. Ahmed, J. Newman, M. Sahu, D. Robinson, L. Siam, C. B. Reamer, T. Zhan, M. Goldstein, S.
Kurtz, M. R. Moster, L. A. Hark, L. J. Katz, “The Effect of an Innovative Vision Simulator (OrCam) on Quality of
Life in Patients with Glaucoma,” Journal of Visual Impariment & Blindness, Vo. 113, No. 4, pp. 332-340, 2019.
[22] Korea Employment Development Institute, “2017 Disabled Statictics (Korean),” 2017.
[23] WHO, “Global data on visual impairments,” 2010.
[24] L. Li and C. L. Tan, “Character Recognition under Severe Perspective Distortion,” Proceedings of IEEE 19th
International Conference on Pattern Recognition, ICPR2008. Tampa, USA, 2008.
[25] K. Wang, et al., “End-to-end scene text recognition,” Proceedings of International Conference on Computer Vision,
ICCV2011, Barcelona, Spain, 2011.
[26] Z. Zhang, “Microsoft Kinect sensor and its effect,” IEEE MultiMedia, Vol. 19, No. 2, pp. 4-10, 2012.
BIOGRAPHIES OF AUTHORS
Jaejoon Kim received the MS and Ph.D. degree from the Department of Electrical Engineering,
Iowa State University, USA. He received Bachelors from Department Electronics Engineering
and Mathematics, Hanyang University, Korea. From 2001 to 2002, he was a senior researcher at
ETRI (Electronics and Telecommunication Research Institute). He is currently a professor at
Daegu University, Republic of Korea. His research interests include image processing, neural
networks and non-destructive evaluation.

Contenu connexe

Tendances

Smartcard based Android Application for Public Transport Ticketing System
Smartcard based Android Application for Public Transport Ticketing SystemSmartcard based Android Application for Public Transport Ticketing System
Smartcard based Android Application for Public Transport Ticketing System
Neelam Gulrajani
 
Godeye An Efficient System for Blinds
Godeye An Efficient System for BlindsGodeye An Efficient System for Blinds
Godeye An Efficient System for Blinds
ijtsrd
 

Tendances (20)

IRJET - For(E)Sight :A Perceptive Device to Assist Blind People
IRJET -  	  For(E)Sight :A Perceptive Device to Assist Blind PeopleIRJET -  	  For(E)Sight :A Perceptive Device to Assist Blind People
IRJET - For(E)Sight :A Perceptive Device to Assist Blind People
 
CONTEXT-BASED BARRIER NOTIFICATION SERVICE TOWARD OUTDOOR SUPPORT FOR THE ELD...
CONTEXT-BASED BARRIER NOTIFICATION SERVICE TOWARD OUTDOOR SUPPORT FOR THE ELD...CONTEXT-BASED BARRIER NOTIFICATION SERVICE TOWARD OUTDOOR SUPPORT FOR THE ELD...
CONTEXT-BASED BARRIER NOTIFICATION SERVICE TOWARD OUTDOOR SUPPORT FOR THE ELD...
 
IRJET- Review on Raspberry Pi based Assistive Communication System for Blind,...
IRJET- Review on Raspberry Pi based Assistive Communication System for Blind,...IRJET- Review on Raspberry Pi based Assistive Communication System for Blind,...
IRJET- Review on Raspberry Pi based Assistive Communication System for Blind,...
 
Smartcard based Android Application for Public Transport Ticketing System
Smartcard based Android Application for Public Transport Ticketing SystemSmartcard based Android Application for Public Transport Ticketing System
Smartcard based Android Application for Public Transport Ticketing System
 
Face recognition smart cane using haar-like features and eigenfaces
Face recognition smart cane using haar-like features and eigenfacesFace recognition smart cane using haar-like features and eigenfaces
Face recognition smart cane using haar-like features and eigenfaces
 
Providing Accident Detection in Vehicular Networks through OBD-II Devices and...
Providing Accident Detection in Vehicular Networks through OBD-II Devices and...Providing Accident Detection in Vehicular Networks through OBD-II Devices and...
Providing Accident Detection in Vehicular Networks through OBD-II Devices and...
 
IRJET- Healthy Beat
IRJET- Healthy BeatIRJET- Healthy Beat
IRJET- Healthy Beat
 
IRJET- Design and Development of Tesseract-OCR Based Assistive System to Conv...
IRJET- Design and Development of Tesseract-OCR Based Assistive System to Conv...IRJET- Design and Development of Tesseract-OCR Based Assistive System to Conv...
IRJET- Design and Development of Tesseract-OCR Based Assistive System to Conv...
 
Detection of immovable objects on visually impaired people walking aids
Detection of immovable objects on visually impaired people walking aidsDetection of immovable objects on visually impaired people walking aids
Detection of immovable objects on visually impaired people walking aids
 
Communication among blind, deaf and dumb People
Communication among blind, deaf and dumb PeopleCommunication among blind, deaf and dumb People
Communication among blind, deaf and dumb People
 
Shravan
ShravanShravan
Shravan
 
SMARCOS Abstract Paper submitted to ICCHP 2012
SMARCOS Abstract Paper submitted to ICCHP 2012SMARCOS Abstract Paper submitted to ICCHP 2012
SMARCOS Abstract Paper submitted to ICCHP 2012
 
Smart Voting System with Face Recognition
Smart Voting System with Face RecognitionSmart Voting System with Face Recognition
Smart Voting System with Face Recognition
 
IRJET- Voice Assistant for Visually Impaired People
IRJET-  	  Voice Assistant for Visually Impaired PeopleIRJET-  	  Voice Assistant for Visually Impaired People
IRJET- Voice Assistant for Visually Impaired People
 
IRJET - Sign Language Text to Speech Converter using Image Processing and...
IRJET -  	  Sign Language Text to Speech Converter using Image Processing and...IRJET -  	  Sign Language Text to Speech Converter using Image Processing and...
IRJET - Sign Language Text to Speech Converter using Image Processing and...
 
Godeye An Efficient System for Blinds
Godeye An Efficient System for BlindsGodeye An Efficient System for Blinds
Godeye An Efficient System for Blinds
 
Face Recognition and Door Opening Assistant for Visually Impaired
Face Recognition and Door Opening Assistant for Visually ImpairedFace Recognition and Door Opening Assistant for Visually Impaired
Face Recognition and Door Opening Assistant for Visually Impaired
 
IRJET - Expiry Date and Cost Tracking in Medicine for Visually Impaired
IRJET - Expiry Date and Cost Tracking in Medicine for Visually ImpairedIRJET - Expiry Date and Cost Tracking in Medicine for Visually Impaired
IRJET - Expiry Date and Cost Tracking in Medicine for Visually Impaired
 
IRJET- Development of a Face Recognition System with Deep Learning and Py...
IRJET-  	  Development of a Face Recognition System with Deep Learning and Py...IRJET-  	  Development of a Face Recognition System with Deep Learning and Py...
IRJET- Development of a Face Recognition System with Deep Learning and Py...
 
Survey on Human Computer interaction for disabled persons
Survey on Human Computer interaction  for disabled persons Survey on Human Computer interaction  for disabled persons
Survey on Human Computer interaction for disabled persons
 

Similaire à Application on character recognition system on road sign for visually impaired: case study approach and future

Paper_39-SRAVIP_Smart_Robot_Assistant_for_Visually_Impaired.pdf
Paper_39-SRAVIP_Smart_Robot_Assistant_for_Visually_Impaired.pdfPaper_39-SRAVIP_Smart_Robot_Assistant_for_Visually_Impaired.pdf
Paper_39-SRAVIP_Smart_Robot_Assistant_for_Visually_Impaired.pdf
SathishRadcliffe
 

Similaire à Application on character recognition system on road sign for visually impaired: case study approach and future (20)

An Assistive System for Visually Impaired People
An Assistive System for Visually Impaired PeopleAn Assistive System for Visually Impaired People
An Assistive System for Visually Impaired People
 
IRJET- Navigation and Camera Reading System for Visually Impaired
IRJET- Navigation and Camera Reading System for Visually ImpairedIRJET- Navigation and Camera Reading System for Visually Impaired
IRJET- Navigation and Camera Reading System for Visually Impaired
 
Forey: An Android Application for the Visually Impaired
Forey: An Android Application for the Visually ImpairedForey: An Android Application for the Visually Impaired
Forey: An Android Application for the Visually Impaired
 
Eye-Blink Detection System for Virtual Keyboard
Eye-Blink Detection System for Virtual KeyboardEye-Blink Detection System for Virtual Keyboard
Eye-Blink Detection System for Virtual Keyboard
 
Visual Product Identification For Blind Peoples
Visual Product Identification For Blind PeoplesVisual Product Identification For Blind Peoples
Visual Product Identification For Blind Peoples
 
Sign Language Recognition
Sign Language RecognitionSign Language Recognition
Sign Language Recognition
 
Text Detection and Recognition with Speech Output for Visually Challenged Per...
Text Detection and Recognition with Speech Output for Visually Challenged Per...Text Detection and Recognition with Speech Output for Visually Challenged Per...
Text Detection and Recognition with Speech Output for Visually Challenged Per...
 
Paper_39-SRAVIP_Smart_Robot_Assistant_for_Visually_Impaired.pdf
Paper_39-SRAVIP_Smart_Robot_Assistant_for_Visually_Impaired.pdfPaper_39-SRAVIP_Smart_Robot_Assistant_for_Visually_Impaired.pdf
Paper_39-SRAVIP_Smart_Robot_Assistant_for_Visually_Impaired.pdf
 
Product Label Reading System for visually challenged people
Product Label Reading System for visually challenged peopleProduct Label Reading System for visually challenged people
Product Label Reading System for visually challenged people
 
Sign Language Recognition using Mediapipe
Sign Language Recognition using MediapipeSign Language Recognition using Mediapipe
Sign Language Recognition using Mediapipe
 
F2R Analyzer Using Machine Learning and Deep Learning
F2R Analyzer Using Machine Learning and Deep LearningF2R Analyzer Using Machine Learning and Deep Learning
F2R Analyzer Using Machine Learning and Deep Learning
 
IRJET - Smart E – Cane for the Visually Challenged and Blind using ML Con...
IRJET -  	  Smart E – Cane for the Visually Challenged and Blind using ML Con...IRJET -  	  Smart E – Cane for the Visually Challenged and Blind using ML Con...
IRJET - Smart E – Cane for the Visually Challenged and Blind using ML Con...
 
IRJET- Persons Identification Tool for Visually Impaired - Digital Eye
IRJET-  	  Persons Identification Tool for Visually Impaired - Digital EyeIRJET-  	  Persons Identification Tool for Visually Impaired - Digital Eye
IRJET- Persons Identification Tool for Visually Impaired - Digital Eye
 
IRJET- Review on Text Recognization of Product for Blind Person using MATLAB
IRJET-  Review on Text Recognization of Product for Blind Person using MATLABIRJET-  Review on Text Recognization of Product for Blind Person using MATLAB
IRJET- Review on Text Recognization of Product for Blind Person using MATLAB
 
IRJET- Hand Gesture Recognition System using Convolutional Neural Networks
IRJET- Hand Gesture Recognition System using Convolutional Neural NetworksIRJET- Hand Gesture Recognition System using Convolutional Neural Networks
IRJET- Hand Gesture Recognition System using Convolutional Neural Networks
 
IRJET-Human Face Detection and Identification using Deep Metric Learning
IRJET-Human Face Detection and Identification using Deep Metric LearningIRJET-Human Face Detection and Identification using Deep Metric Learning
IRJET-Human Face Detection and Identification using Deep Metric Learning
 
IRJET- A Smart Personal AI Assistant for Visually Impaired People: A Survey
IRJET-  	  A Smart Personal AI Assistant for Visually Impaired People: A SurveyIRJET-  	  A Smart Personal AI Assistant for Visually Impaired People: A Survey
IRJET- A Smart Personal AI Assistant for Visually Impaired People: A Survey
 
IRJET- VI Spectacle – A Visual Aid for the Visually Impaired
IRJET- VI Spectacle – A Visual Aid for the Visually ImpairedIRJET- VI Spectacle – A Visual Aid for the Visually Impaired
IRJET- VI Spectacle – A Visual Aid for the Visually Impaired
 
DHWANI- THE VOICE OF DEAF AND MUTE
DHWANI- THE VOICE OF DEAF AND MUTEDHWANI- THE VOICE OF DEAF AND MUTE
DHWANI- THE VOICE OF DEAF AND MUTE
 
DHWANI- THE VOICE OF DEAF AND MUTE
DHWANI- THE VOICE OF DEAF AND MUTEDHWANI- THE VOICE OF DEAF AND MUTE
DHWANI- THE VOICE OF DEAF AND MUTE
 

Plus de IJECEIAES

Predicting churn with filter-based techniques and deep learning
Predicting churn with filter-based techniques and deep learningPredicting churn with filter-based techniques and deep learning
Predicting churn with filter-based techniques and deep learning
IJECEIAES
 
Automatic customer review summarization using deep learningbased hybrid senti...
Automatic customer review summarization using deep learningbased hybrid senti...Automatic customer review summarization using deep learningbased hybrid senti...
Automatic customer review summarization using deep learningbased hybrid senti...
IJECEIAES
 
Situational judgment test measures administrator computational thinking with ...
Situational judgment test measures administrator computational thinking with ...Situational judgment test measures administrator computational thinking with ...
Situational judgment test measures administrator computational thinking with ...
IJECEIAES
 
Hand LightWeightNet: an optimized hand pose estimation for interactive mobile...
Hand LightWeightNet: an optimized hand pose estimation for interactive mobile...Hand LightWeightNet: an optimized hand pose estimation for interactive mobile...
Hand LightWeightNet: an optimized hand pose estimation for interactive mobile...
IJECEIAES
 
Application of deep learning methods for automated analysis of retinal struct...
Application of deep learning methods for automated analysis of retinal struct...Application of deep learning methods for automated analysis of retinal struct...
Application of deep learning methods for automated analysis of retinal struct...
IJECEIAES
 
Predictive modeling for breast cancer based on machine learning algorithms an...
Predictive modeling for breast cancer based on machine learning algorithms an...Predictive modeling for breast cancer based on machine learning algorithms an...
Predictive modeling for breast cancer based on machine learning algorithms an...
IJECEIAES
 

Plus de IJECEIAES (20)

Predicting churn with filter-based techniques and deep learning
Predicting churn with filter-based techniques and deep learningPredicting churn with filter-based techniques and deep learning
Predicting churn with filter-based techniques and deep learning
 
Taxi-out time prediction at Mohammed V Casablanca Airport
Taxi-out time prediction at Mohammed V Casablanca AirportTaxi-out time prediction at Mohammed V Casablanca Airport
Taxi-out time prediction at Mohammed V Casablanca Airport
 
Automatic customer review summarization using deep learningbased hybrid senti...
Automatic customer review summarization using deep learningbased hybrid senti...Automatic customer review summarization using deep learningbased hybrid senti...
Automatic customer review summarization using deep learningbased hybrid senti...
 
Ataxic person prediction using feature optimized based on machine learning model
Ataxic person prediction using feature optimized based on machine learning modelAtaxic person prediction using feature optimized based on machine learning model
Ataxic person prediction using feature optimized based on machine learning model
 
Situational judgment test measures administrator computational thinking with ...
Situational judgment test measures administrator computational thinking with ...Situational judgment test measures administrator computational thinking with ...
Situational judgment test measures administrator computational thinking with ...
 
Hand LightWeightNet: an optimized hand pose estimation for interactive mobile...
Hand LightWeightNet: an optimized hand pose estimation for interactive mobile...Hand LightWeightNet: an optimized hand pose estimation for interactive mobile...
Hand LightWeightNet: an optimized hand pose estimation for interactive mobile...
 
An overlapping conscious relief-based feature subset selection method
An overlapping conscious relief-based feature subset selection methodAn overlapping conscious relief-based feature subset selection method
An overlapping conscious relief-based feature subset selection method
 
Recognition of music symbol notation using convolutional neural network
Recognition of music symbol notation using convolutional neural networkRecognition of music symbol notation using convolutional neural network
Recognition of music symbol notation using convolutional neural network
 
A simplified classification computational model of opinion mining using deep ...
A simplified classification computational model of opinion mining using deep ...A simplified classification computational model of opinion mining using deep ...
A simplified classification computational model of opinion mining using deep ...
 
An efficient convolutional neural network-extreme gradient boosting hybrid de...
An efficient convolutional neural network-extreme gradient boosting hybrid de...An efficient convolutional neural network-extreme gradient boosting hybrid de...
An efficient convolutional neural network-extreme gradient boosting hybrid de...
 
Generating images using generative adversarial networks based on text descrip...
Generating images using generative adversarial networks based on text descrip...Generating images using generative adversarial networks based on text descrip...
Generating images using generative adversarial networks based on text descrip...
 
Collusion-resistant multiparty data sharing in social networks
Collusion-resistant multiparty data sharing in social networksCollusion-resistant multiparty data sharing in social networks
Collusion-resistant multiparty data sharing in social networks
 
Application of deep learning methods for automated analysis of retinal struct...
Application of deep learning methods for automated analysis of retinal struct...Application of deep learning methods for automated analysis of retinal struct...
Application of deep learning methods for automated analysis of retinal struct...
 
Proposal of a similarity measure for unified modeling language class diagram ...
Proposal of a similarity measure for unified modeling language class diagram ...Proposal of a similarity measure for unified modeling language class diagram ...
Proposal of a similarity measure for unified modeling language class diagram ...
 
Efficient criticality oriented service brokering policy in cloud datacenters
Efficient criticality oriented service brokering policy in cloud datacentersEfficient criticality oriented service brokering policy in cloud datacenters
Efficient criticality oriented service brokering policy in cloud datacenters
 
System call frequency analysis-based generative adversarial network model for...
System call frequency analysis-based generative adversarial network model for...System call frequency analysis-based generative adversarial network model for...
System call frequency analysis-based generative adversarial network model for...
 
Comparison of Iris dataset classification with Gaussian naïve Bayes and decis...
Comparison of Iris dataset classification with Gaussian naïve Bayes and decis...Comparison of Iris dataset classification with Gaussian naïve Bayes and decis...
Comparison of Iris dataset classification with Gaussian naïve Bayes and decis...
 
Efficient intelligent crawler for hamming distance based on prioritization of...
Efficient intelligent crawler for hamming distance based on prioritization of...Efficient intelligent crawler for hamming distance based on prioritization of...
Efficient intelligent crawler for hamming distance based on prioritization of...
 
Predictive modeling for breast cancer based on machine learning algorithms an...
Predictive modeling for breast cancer based on machine learning algorithms an...Predictive modeling for breast cancer based on machine learning algorithms an...
Predictive modeling for breast cancer based on machine learning algorithms an...
 
An algorithm for decomposing variations of 3D model
An algorithm for decomposing variations of 3D modelAn algorithm for decomposing variations of 3D model
An algorithm for decomposing variations of 3D model
 

Dernier

AKTU Computer Networks notes --- Unit 3.pdf
AKTU Computer Networks notes ---  Unit 3.pdfAKTU Computer Networks notes ---  Unit 3.pdf
AKTU Computer Networks notes --- Unit 3.pdf
ankushspencer015
 
Call for Papers - Educational Administration: Theory and Practice, E-ISSN: 21...
Call for Papers - Educational Administration: Theory and Practice, E-ISSN: 21...Call for Papers - Educational Administration: Theory and Practice, E-ISSN: 21...
Call for Papers - Educational Administration: Theory and Practice, E-ISSN: 21...
Christo Ananth
 
Call for Papers - African Journal of Biological Sciences, E-ISSN: 2663-2187, ...
Call for Papers - African Journal of Biological Sciences, E-ISSN: 2663-2187, ...Call for Papers - African Journal of Biological Sciences, E-ISSN: 2663-2187, ...
Call for Papers - African Journal of Biological Sciences, E-ISSN: 2663-2187, ...
Christo Ananth
 
FULL ENJOY Call Girls In Mahipalpur Delhi Contact Us 8377877756
FULL ENJOY Call Girls In Mahipalpur Delhi Contact Us 8377877756FULL ENJOY Call Girls In Mahipalpur Delhi Contact Us 8377877756
FULL ENJOY Call Girls In Mahipalpur Delhi Contact Us 8377877756
dollysharma2066
 
Call Now ≽ 9953056974 ≼🔝 Call Girls In New Ashok Nagar ≼🔝 Delhi door step de...
Call Now ≽ 9953056974 ≼🔝 Call Girls In New Ashok Nagar  ≼🔝 Delhi door step de...Call Now ≽ 9953056974 ≼🔝 Call Girls In New Ashok Nagar  ≼🔝 Delhi door step de...
Call Now ≽ 9953056974 ≼🔝 Call Girls In New Ashok Nagar ≼🔝 Delhi door step de...
9953056974 Low Rate Call Girls In Saket, Delhi NCR
 

Dernier (20)

Unleashing the Power of the SORA AI lastest leap
Unleashing the Power of the SORA AI lastest leapUnleashing the Power of the SORA AI lastest leap
Unleashing the Power of the SORA AI lastest leap
 
CCS335 _ Neural Networks and Deep Learning Laboratory_Lab Complete Record
CCS335 _ Neural Networks and Deep Learning Laboratory_Lab Complete RecordCCS335 _ Neural Networks and Deep Learning Laboratory_Lab Complete Record
CCS335 _ Neural Networks and Deep Learning Laboratory_Lab Complete Record
 
(INDIRA) Call Girl Bhosari Call Now 8617697112 Bhosari Escorts 24x7
(INDIRA) Call Girl Bhosari Call Now 8617697112 Bhosari Escorts 24x7(INDIRA) Call Girl Bhosari Call Now 8617697112 Bhosari Escorts 24x7
(INDIRA) Call Girl Bhosari Call Now 8617697112 Bhosari Escorts 24x7
 
Bhosari ( Call Girls ) Pune 6297143586 Hot Model With Sexy Bhabi Ready For ...
Bhosari ( Call Girls ) Pune  6297143586  Hot Model With Sexy Bhabi Ready For ...Bhosari ( Call Girls ) Pune  6297143586  Hot Model With Sexy Bhabi Ready For ...
Bhosari ( Call Girls ) Pune 6297143586 Hot Model With Sexy Bhabi Ready For ...
 
Call Girls Wakad Call Me 7737669865 Budget Friendly No Advance Booking
Call Girls Wakad Call Me 7737669865 Budget Friendly No Advance BookingCall Girls Wakad Call Me 7737669865 Budget Friendly No Advance Booking
Call Girls Wakad Call Me 7737669865 Budget Friendly No Advance Booking
 
chapter 5.pptx: drainage and irrigation engineering
chapter 5.pptx: drainage and irrigation engineeringchapter 5.pptx: drainage and irrigation engineering
chapter 5.pptx: drainage and irrigation engineering
 
AKTU Computer Networks notes --- Unit 3.pdf
AKTU Computer Networks notes ---  Unit 3.pdfAKTU Computer Networks notes ---  Unit 3.pdf
AKTU Computer Networks notes --- Unit 3.pdf
 
Booking open Available Pune Call Girls Pargaon 6297143586 Call Hot Indian Gi...
Booking open Available Pune Call Girls Pargaon  6297143586 Call Hot Indian Gi...Booking open Available Pune Call Girls Pargaon  6297143586 Call Hot Indian Gi...
Booking open Available Pune Call Girls Pargaon 6297143586 Call Hot Indian Gi...
 
Call for Papers - Educational Administration: Theory and Practice, E-ISSN: 21...
Call for Papers - Educational Administration: Theory and Practice, E-ISSN: 21...Call for Papers - Educational Administration: Theory and Practice, E-ISSN: 21...
Call for Papers - Educational Administration: Theory and Practice, E-ISSN: 21...
 
The Most Attractive Pune Call Girls Manchar 8250192130 Will You Miss This Cha...
The Most Attractive Pune Call Girls Manchar 8250192130 Will You Miss This Cha...The Most Attractive Pune Call Girls Manchar 8250192130 Will You Miss This Cha...
The Most Attractive Pune Call Girls Manchar 8250192130 Will You Miss This Cha...
 
Call Girls Pimpri Chinchwad Call Me 7737669865 Budget Friendly No Advance Boo...
Call Girls Pimpri Chinchwad Call Me 7737669865 Budget Friendly No Advance Boo...Call Girls Pimpri Chinchwad Call Me 7737669865 Budget Friendly No Advance Boo...
Call Girls Pimpri Chinchwad Call Me 7737669865 Budget Friendly No Advance Boo...
 
Double rodded leveling 1 pdf activity 01
Double rodded leveling 1 pdf activity 01Double rodded leveling 1 pdf activity 01
Double rodded leveling 1 pdf activity 01
 
BSides Seattle 2024 - Stopping Ethan Hunt From Taking Your Data.pptx
BSides Seattle 2024 - Stopping Ethan Hunt From Taking Your Data.pptxBSides Seattle 2024 - Stopping Ethan Hunt From Taking Your Data.pptx
BSides Seattle 2024 - Stopping Ethan Hunt From Taking Your Data.pptx
 
Call for Papers - African Journal of Biological Sciences, E-ISSN: 2663-2187, ...
Call for Papers - African Journal of Biological Sciences, E-ISSN: 2663-2187, ...Call for Papers - African Journal of Biological Sciences, E-ISSN: 2663-2187, ...
Call for Papers - African Journal of Biological Sciences, E-ISSN: 2663-2187, ...
 
Thermal Engineering -unit - III & IV.ppt
Thermal Engineering -unit - III & IV.pptThermal Engineering -unit - III & IV.ppt
Thermal Engineering -unit - III & IV.ppt
 
Intze Overhead Water Tank Design by Working Stress - IS Method.pdf
Intze Overhead Water Tank  Design by Working Stress - IS Method.pdfIntze Overhead Water Tank  Design by Working Stress - IS Method.pdf
Intze Overhead Water Tank Design by Working Stress - IS Method.pdf
 
FULL ENJOY Call Girls In Mahipalpur Delhi Contact Us 8377877756
FULL ENJOY Call Girls In Mahipalpur Delhi Contact Us 8377877756FULL ENJOY Call Girls In Mahipalpur Delhi Contact Us 8377877756
FULL ENJOY Call Girls In Mahipalpur Delhi Contact Us 8377877756
 
(INDIRA) Call Girl Meerut Call Now 8617697112 Meerut Escorts 24x7
(INDIRA) Call Girl Meerut Call Now 8617697112 Meerut Escorts 24x7(INDIRA) Call Girl Meerut Call Now 8617697112 Meerut Escorts 24x7
(INDIRA) Call Girl Meerut Call Now 8617697112 Meerut Escorts 24x7
 
Call Girls Walvekar Nagar Call Me 7737669865 Budget Friendly No Advance Booking
Call Girls Walvekar Nagar Call Me 7737669865 Budget Friendly No Advance BookingCall Girls Walvekar Nagar Call Me 7737669865 Budget Friendly No Advance Booking
Call Girls Walvekar Nagar Call Me 7737669865 Budget Friendly No Advance Booking
 
Call Now ≽ 9953056974 ≼🔝 Call Girls In New Ashok Nagar ≼🔝 Delhi door step de...
Call Now ≽ 9953056974 ≼🔝 Call Girls In New Ashok Nagar  ≼🔝 Delhi door step de...Call Now ≽ 9953056974 ≼🔝 Call Girls In New Ashok Nagar  ≼🔝 Delhi door step de...
Call Now ≽ 9953056974 ≼🔝 Call Girls In New Ashok Nagar ≼🔝 Delhi door step de...
 

Application on character recognition system on road sign for visually impaired: case study approach and future

  • 1. International Journal of Electrical and Computer Engineering (IJECE) Vol. 10, No. 1, February 2020, pp. 778~785 ISSN: 2088-8708, DOI: 10.11591/ijece.v10i1.pp778-785  778 Journal homepage: http://ijece.iaescore.com/index.php/IJECE Application on character recognition system on road sign for visually impaired: case study approach and future Jaejoon Kim School of Computer and Communication, Daegu University, South Korea Article Info ABSTRACT Article history: Received Apr 7, 2019 Revised Sep 18, 2019 Accepted Sep 27, 2019 Many visually impaired people worldwide are unable to travel safely and autonomously because they are physically unable to perceive effective visual information during their daily lives. In this research, we study how to extract the character information of the road sign and transmit it to the visually impaired effectively, so they can understand easier. Experimental method is to apply the Maximally Stable External Region and Stroke Width Transform method in Phase I so that the visually impaired person can recognize the letters on the road signs. It is to convey text information to the disabled. The result of Phase I using samples of simple road signs was to extract the sign information after dividing the exact character area, but the accuracy was not good for the Hangul (Korean characters) information. The initial experimental results in the Phase II succeeded in transmitting the text information on Phase I to the visually impaired. In the future, it will be required to develop a wearable character recognition system that can be attached to the visually impaired. In order to perform this task, we need to develop and verify a miniaturized and wearable character recognition system. In this paper, we examined the method of recognizing road sign characters on the road and presented a possibility that may be applicable to our final development. Keywords: Character recognition system MSER OCR Visually impaired Wearable device Copyright © 2020 Institute of Advanced Engineering and Science. All rights reserved. Corresponding Author: Jaejoon Kim School of Computer and Communication, Daegu University, 201 Daegudae-ro, Gyeongsan, Gyeongbuk, 38453, South Korea. Email: jjkimisu@daegu.ac.kr 1. INTRODUCTION There are many people with visual disabilities in the world as well as people reside in Korea. Auxiliary devices have been actively developed to provide visually impaired persons with accurate information on road signs during walking, which can be proven as crucial parts of everyday social activities. Even more, various devices such as products that read or provide directions for blind have been actively developed in the past. Various methods are currently in development to automatically extract the region surrounding a character when that character is detected to be in the presence of a visually impaired person [1-5]. The method consists of mainly first recognizing the character string in its region and then sending the scanned information back to the user. Due to artificial intelligence technology, additional features can include a dictionary function in case the user does not understand the meaning of the sign in question. In recent years, various techniques using artificial intelligence in image recognition, natural language processing, and natural language generation have been developed to help people with visual impairments to maintain a comprehensive and productive life [6].
  • 2. Int J Elec & Comp Eng ISSN: 2088-8708  Application on character recognition system on road sign for visually impaired: case ... (Jaejoon Kim) 779 In this paper, we tried to develop a variety of wearable assistive aids for the visually impaired to eliminate the inconvenience of walking or real life. In order to do this, in Phase I, Maximally Stable External Region (MSER) and Stroke Width Transform (SWT) features were applied to character recognition of road signs and pedestrian signs [7, 8]. In Phase II, we tried to design a text recognition system based on TTS (Text-to-Speech) converter [9] that transmits the text recognition information to the visually impaired. This paper proposes a plan for the visually impaired to meet these developments and to present the applicable possibilities. Section 2 briefly discusses the related works for character string recognition technology currently in development for the visually impaired and characteristics of information guiding system. Section 3 describes experimental implementation of character recognition for our proposed system. Sections 4 and 5 conclude the paper by discussing experimental results, future works and other possible solutions. 2. RELATED WORK 2.1. Characteristics and trend of information guiding system Blind people are divided into low vision and the blind. Low vision refers to vision problems that cannot be improved by medical or optical means due to birth defects or acquired eye diseases. According to a report of Yonsei University Medical Center, the suicide rate of the low vision is twice that of the normal person, and the psychological stress due to the low vision is large, and the need for an auxiliary device to help the low vision is increasing [10]. In addition, a report from the UN estimates that the world‟s population age 60 or older will increase from 11% in 2012 to 22% by 2020 as shown in Figure 1. Especially in North America and Europe, the distribution of population aged 60 years or older is reported to be significantly higher than the average [11]. In the point of world market, Figure 2 indicated that the global elderly and disabled assistive devices market was valued at US$ 14,109.1 million in 2015 and is expected to expand at a CAGR of 7.4% during the forecast period (2016 – 2024) [12]. Figure 1. The statistcs of UN, gendered innovations (2012 vs. 2050) [11] Figure 2. Global elderly and disabled assistive devices market size and forecast (unit: US$ Million) [12]
  • 3.  ISSN: 2088-8708 Int J Elec & Comp Eng, Vol. 10, No. 1, February 2020 : 778 - 785 780 2.2. Character information recognition system To allow a visually impared person to recognize a character, our application must be able to find the character region in which the sign or the message is written. The most well-known method used for binarization is a global binarization which is based on a single threshold [13-15]. However, these methods have chracteristics such as performance issues with the overall brightness of the input document image are not constant. Another method used for binarization is the Maximally Stable External Region (MSER) algorithm, which is used in a more robust manner of finding a region with a small rate of change while changing the threshold value in order to distinguish characters in both bright and dark regions. The character area is generally easy to distinguish from the background and has a constant brightness value. This characteristic of the character region is suitable to be detected by the MSER method [7, 8]. Various applications have been developed by various companies, all of which attempt to extract a character string by extracting horizontal and vertical boundary components from the input image. These methods basically either find the portion with the greatest boundary component or extract a character string that is irrespective of misjudgment or distortion. The ABBYY FineReader Engine, OCR SDK (character recognition development toolkit) supplied by Retia, was applied to the development of a 'print-to-speech converter', which is a device for recognizing printed characters for visually impaired and converting them into voice [16-18]. ABBYY [19] developed ABBYY FineReader Engine with a goal of a stable „print-to-speech‟ converter for the visually impaired. The OCR key technologies of ABBYY FineReader Engine have several main features such as: a. Korean, Chinese, English, numbers and special characters support b. Vertical reading, Multi reading c. Batch scanning, batch conversion possible d. It can save word, text file etc. By utilizing this engine, the visually impaired user can convert the desired document to speech through this device. While in the past books, receipts and other readable documents could not be read without Braille documents, the engine now makes it possible to directly scan a desired document or image by using the „sound‟ with OCR technology and convert it so the user can hear the by voice. The „finger reader‟ [20] product developed by MIT works by placing a finger near the book. A high-resolution camera is used to scan the characters corresponding to the finger‟s touch while also reading them out loud after being scanned. After several years of research, OrCam, an Israeli company, is now developing new products that use facial recognition technology to scan not only characters but also various colors [21]. There are also „smart glasses‟ for people with partial vision loss. These „smart glasses‟ work by viewing images in 3D glasses. However even though these glasses have been proven useful reducing inconvenience of the visually disabled, they are still very expensive. Therefore there will be difficulties in making them commercially available for everyone. Figure 3 shows the number of people in Korea with sight disabilities is around about 250,000 people as of 2016. Unfortunately, the number of people with visual disabilities is growing larger each year by more than 10%. In addition, the types of disabilities are discribed in Figure 4. Figure 3. The statistics of disabilities in Korea, 2007 and 2016 [22]
  • 4. Int J Elec & Comp Eng ISSN: 2088-8708  Application on character recognition system on road sign for visually impaired: case ... (Jaejoon Kim) 781 Figure 4. The types of visually impaired. (a) Scenery seen with normal eye, (b) Scenery seen with eyes with cataracts, (c) Scenery seen with eyes with glaucoma, (d) Scenery seen with eyes with diabetic retinopathy, (e) Scenery seen with eyes with macular degeneration [23] 3. EXPERIMENTAL IMPLEMENTATION One of the most important parts of the application is to provide information that is only necessary in order to never confuse the user. Therefore, the application must be able to select only the relevant area amongst the other „background noise‟. For example, the application must be able to distinguish between a road sign and the other natural background surrounding the sign. In this paper, we aim to detect the text area in the image provided with the visually impaired and to recognize the information provided in the area as a character and deliver it to the visually impaired. In order to implement our goal, we used MSER and OCR methods for detecting information area and text detection as a starting point of research. Based on the Phase I of the research, we aim to develop a wearable device that provides road information about the visually impaired by utilizing the TTS (Text-to-Speech) function of a typical smartphone. 3.1. Feature extraction with MSER Information on the road signs do not contain as much textual information when compared to books or newspapers. As a method of the visually impaired to arrive at their desired destination, they had to previous rely the usage of a guidance stick. While it may to practical and has been used as from the beginning of blind people, a different method can be developed. These days it is possible to detect the text in an image using no more than a smartphone with a camera. By using pixel spans as nodes, blobs can be generated which can detect text in an image more quickly and accurately. In addition, as the usage of smartphone increased very rapidly, an optical character recognition application that recognizes a character of an image captured by a camera mounted on a smart phone and displays it on a display can be distributed in the near future and prove more practical compared to a guidance stick. The most important step is to detect the blob which does a candidate region exists in the image that is most likely to contain the text of the road sign. The SWT algorithm [24, 25] is a method of detecting text by determining which region has little change in the shape and thickness. When the parameters are given, the algorithm detects the text within the image using those parameters. The main advantage of SWT are that it can detect text in an image without a separate learning process, but it also has a drawback in which it takes a lot of time to detect text in an image because of its complicated operation. The MSER algorithm for robust text detection is widely used as an algorithm for detecting blob as an aggregate region of pixels different from surrounding pixels in intensity. The MSER algorithm has the advantage that it can detect the blob faster than the SWT algorithm, but it has a disadvantage that the accuracy is somewhat lacking, such as detecting the tickle or noise by blob. 3.2. Application procedure Figure 5 is an overall block diagram that shows the process of our application after capturing the image and extracting the text. In this paper, we aimed to test the design possibility of an assistive device which can be helpful for the visually impaired through Phase I process in two stages. The first step captures images of a smartphone or future terminal device that the blind has and then removes non-text regions using MSER. The MSER extraction implements the following steps:
  • 5.  ISSN: 2088-8708 Int J Elec & Comp Eng, Vol. 10, No. 1, February 2020 : 778 - 785 782 Figure 5. The overall block diagram a. Sweep threshold of intensity b. Extract connected components c. Find a threshold on maximally stable d. Approximate a region with an ellipse e. Keep region descriptors One of the most important features of the MSER algorithm is that it detects most text in an image but also extracts a stable region, not a text region. To make the algorithm suit our needs, we removed non- text areas using parameters that distinguish words by their geometrical properties within the image. In addition, SWT can be used to distinguish more sophisticated non-text areas. In this paper, we applied this method to find and remove non-text areas. In order to merge individual text areas to extract information on a single word or a text line after finally confirming the text area, a neighboring text area is searched to form a bounding box of these areas, and finally, the information is extracted. In Phase II, the final detected text information is converted into voice information on the visually impaired to be transmitted or applied to future applicable and wearable devices. 4. RESULT AND DISCUSSION 4.1. Phase I In our experiment, we used a typical image of the road sign which the visually impaired can easily encounter on the road. The left column of Figure 6 shows test data set. These images were sorted in order to make test conditions more difficult. When comparing the amount of text to the whole image there is not much text contained in the image in the first place. Therefore, these images were ideal in testing if our proposed method of combining MSER and SWT to extract the necessary information. These test conditions were also important because the end goal of our application is to distribute it so that it can be used during everyday activities. Thus, we made sure that the images were selected very carefully in order to simulate what the user might encounter while using the application. In Figure 6, we show the results of applying MSER, geometric characteristics, and then applying SWT to find non-text areas in captured images using mobile phones of the visually impaired. These results are shown on right column of Figure 6 and Table 1 shows the final detected character information of each test image. In our test case of images containing only English characters, accuracy was high in the implemented method. However, when we applied our method to an image where Korean and English were mixed, we could see a drop off in performance. Through this result we can confirm that a different detection engine is required for the detection of multiple languages.
  • 6. Int J Elec & Comp Eng ISSN: 2088-8708  Application on character recognition system on road sign for visually impaired: case ... (Jaejoon Kim) 783 (a) (b) (c) (d) (e) (f) (g) (h) Figure 6. The test data set and experimental results Table 1. Implemental and detected results Test data Detected Text Information Figure 6-(a) Cyclists AHEAD SLOW Figure 6-(c) AHEAD Figure 6-(e) PED XING Figure 6-(g) ‟.„i".Si6 SLOW 4.2. Phase II TTS is a type of speech synthesis program that makes the contents of computer documents, such as help files and web pages, sound human-readable. TTS can also read image information for people with visual impairments. There are many TTS products on the market, including Read Please 2000, Proverbe Speech Unit, and TextAloud. Lucent and AT & T have their own products called “Text-to-Speech.” In this research, we applied Microsoft Speech API's speech to text functionality in consideration of efficient aspects [26]. Figure 7 shows simplified block diagram for TTS function. This system works well for our test data. (a) (b) (c) Figure 7. The block diagram for text-to-speech process. (a) input image, (b) captured letters on notepad, and (c) speech out with text.txt file
  • 7.  ISSN: 2088-8708 Int J Elec & Comp Eng, Vol. 10, No. 1, February 2020 : 778 - 785 784 5. CONCLUSION In this paper, we proposed a character recognition system for visual impaired persons and added in a voice guidance application implementation process. Our goal is to add a character recognition system for the visually impaired to the smartphone terminal through H/W production. We proposed a method to detect the characters and apply the voice guidance from the images captured in Phase I. Using the features of MSER and SWT, it showed the possibility of extracting the letters of road signs, and in Phase II, it showed the result of delivering the extracted text information to the users. However, in this paper, we can look at various future points to consider. So far, the road guidance service application for the visually impaired has been developed using a blue pitcher communication based H/W auxiliary device such as a beacon in a specific area. With the development of IoT technology, the objects and data related to IoT have been increased greatly. It is necessary to develop a camera-based video information guidance service app that utilizes a voice guidance service technology that provides appropriate information. In addition, artificial intelligence techniques such as image recognition, natural language processing, and natural language generation can be used to enable blind people to live a more comprehensive and productive life. ACKNOWLEDGEMENTS This research was supported by the Daegu University Research Grant, 2017. REFERENCES [1] A. A. Panchal, et al., “Character detection and recognition system for visually impaired people,” Proceedings of IEEE International Conference on Recent Trends in Electronics, Information & Communication Technology, India, pp. 1492-1496, 2016. [2] Abdurrasyid, et al., “Detection of immovable objects on visually impaired people walking aids,” TELKOMNIKA Telecommunication Computing Electronics and Control, vol. 17, pp. 580-585, 2019. [3] D. A. Mohammed, et al., “Off-line handwritten character recognition using an integrated DBSCAN-ANN scheme,” Indonesian Journal of Electrical Engineering and Computer Science, vol. 14, pp. 1443-1451, 2019. [4] O. O. Oladayo, “Yoruba Language and Numerals‟ Offline Interpreter Using Morphological and Template Matching,” IAES International Journal of Artificial Intelligence, vol. 3, pp. 64-72, 2014. [5] I. F. Bt. Hairuman and O. M. Foong, “OCR signage recognition with skew & slant correction for visually impaired people,” Proceedings of IEEE International Conference on Hybrid Intelligent Systems, Melacca, Malaysia, pp. 306- 310, 2011. [6] M. E. Pollack, “Intelligent Technology for an Aging Population: The Use of AI to Assist Elders with Cognitive Impairment,” AI Magazine, American Association for Artificial Intelligence, vol. 26, pp. 9-21, 2005. [7] J. Matas, et al., “Robust Wide Baseline Stereo from Maximally Stable Extremal Regions,” Proceedings of the British Machine Vision Conference, BMVC, pp. 384-393, 2002. [8] Forssen and Per-Erik, “Maximally Stable Colour Regions for Recognition and Matching,” Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, CVPR, Minneapolis, USA, 2007. [9] Q. Ye and D. Doermann, “Text Detection and Recognition in Imagery: A Survey,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 37, pp. 1480-1500, 2015. [10] T. H. Rim, et al., “Influence of visual acuity on suicidal ideation, suicide attempts and depression in South Korea,” British Journal of Ophthalmology, vol. 99, pp. 1112-1119, 2015. [11] Exploring Markets for Assistive Technologies for the Elderly, UN Gendered Innovations. Available: https://genderedinnovations.stanford.edu/case-studies/robots.html. [12] Global Elderly and Disabled Assistive Devices Market, “Coherent Market Insights,” 2017. [13] N. Otsu, “A Threshold Selection Method from Gray Level Histogram,” IEEE Transactions on System, vol. 19, pp. 62-66, 1979. [14] G. Park, et al., “A Study on Enhanced Binarization Method by Using Intensity Information,” Proceedings of the Spring Conference of the Korea Multimedia Society, pp. 441-445, 2003. [15] K. Munadi, et al., “Improved Thresholding Method for Enhancing Jawi Binarization Performance,” Proceedings of 14th IAPR International Conference on Document Analysis and Recognition, Kyoto, Japan, vol. 1, pp. 1108-1113, 2017. [16] R. Smith, “An Overview of the Tesseract OCR Engine,” Proceedings of 9th International Conference on Document Analysis and Recognition, Brazil, 2007. [17] A. Abdulkader and M. R. Cassey, “Low cost correction of OCR errors using learning in a multi-engine environment,” Proceedings of IEEE International Conference on Document Analysis and Recognition, ICDAR2009, pp. 576-580, 2009. [18] T. M. Breuel, et al., “High-performance OCR for printed English and fraktur using LSTM networks,” Proceedings of IEEE International Conference on Document Analysis and Recognition, ICDAR2013. Washington, USA, 2013. [19] M. Heliński, M. Kmieciak, T Parkoła, “Report on the comparison of Tesseract and ABBYY FineReader OCR engines,” Impact, 2012.
  • 8. Int J Elec & Comp Eng ISSN: 2088-8708  Application on character recognition system on road sign for visually impaired: case ... (Jaejoon Kim) 785 [20] R. Shilkrot, J. Huber, W. M. Ee, P. Maes, S. C. Nanayakkara, “FingerReader: A Wearable Device to Explore Printed Text on the Go,” Proceedings of the 33rd Annual ACM Conference on Human Factors in Computing Systems, CHI 2015, Seoul, Korea, 2015. [21] M. Waisbourd, O. M. Ahmed, J. Newman, M. Sahu, D. Robinson, L. Siam, C. B. Reamer, T. Zhan, M. Goldstein, S. Kurtz, M. R. Moster, L. A. Hark, L. J. Katz, “The Effect of an Innovative Vision Simulator (OrCam) on Quality of Life in Patients with Glaucoma,” Journal of Visual Impariment & Blindness, Vo. 113, No. 4, pp. 332-340, 2019. [22] Korea Employment Development Institute, “2017 Disabled Statictics (Korean),” 2017. [23] WHO, “Global data on visual impairments,” 2010. [24] L. Li and C. L. Tan, “Character Recognition under Severe Perspective Distortion,” Proceedings of IEEE 19th International Conference on Pattern Recognition, ICPR2008. Tampa, USA, 2008. [25] K. Wang, et al., “End-to-end scene text recognition,” Proceedings of International Conference on Computer Vision, ICCV2011, Barcelona, Spain, 2011. [26] Z. Zhang, “Microsoft Kinect sensor and its effect,” IEEE MultiMedia, Vol. 19, No. 2, pp. 4-10, 2012. BIOGRAPHIES OF AUTHORS Jaejoon Kim received the MS and Ph.D. degree from the Department of Electrical Engineering, Iowa State University, USA. He received Bachelors from Department Electronics Engineering and Mathematics, Hanyang University, Korea. From 2001 to 2002, he was a senior researcher at ETRI (Electronics and Telecommunication Research Institute). He is currently a professor at Daegu University, Republic of Korea. His research interests include image processing, neural networks and non-destructive evaluation.