SlideShare une entreprise Scribd logo
1  sur  9
Télécharger pour lire hors ligne
International Journal on Natural Language Computing (IJNLC) Vol. 3, No.2, April 2014
DOI : 10.5121/ijnlc.2014.3201 01
AN EXPERT SYSTEM FOR AUTOMATIC READING OF
A TEXT WRITTEN IN STANDARD ARABIC
Tebbi Hanane1
and Azzoune Hamid2
1
LRIA, Option: Representation of Knowledge and systems of Inference, USTHB, Algiers
2
LRIA, Option: Representation of Knowledge and systems of Inference, USTHB,
Algiers
ABSTRACT
In this work we present our expert system of Automatic reading or speech synthesis based on a text
written in Standard Arabic, our work is carried out in two great stages: the creation of the sound data
base, and the transformation of the written text into speech (Text To Speech TTS). This transformation is
done firstly by a Phonetic Orthographical Transcription (POT) of any written Standard Arabic text with
the aim of transforming it into his corresponding phonetics sequence, and secondly by the generation of
the voice signal which corresponds to the chain transcribed. We spread out the different of conception of
the system, as well as the results obtained compared to others works studied to realize TTS based on
Standard Arabic.
KEYWORDS
Engineering knowledge, modelling and representation of vocal knowledge, expert system, (TTS Text To
Speech), Standard Arabic, PRAAT.
1. INTRODUCTION
The automatic speech generation is a complex task because of the variability intra and
interlocutor of the voice signal. In computer science the difficulty of modeling the speech signal
is due to the fact that we don’t know till now how to model very well the enormous mass of
knowledge and information useful for the signal synthesis. So we have made a choice to use an
expert system to model that knowledge to build a robust system which can really read a text
written in a language chosen especially in Standard Arabic. To obtain a better organization of
our work, we defined our direct aims. We divided our modeling into three essential stages; the
signal analysis, the phonetic orthographical transcription (POT) and finally the synthesis of the
textual representation written in Standard Arabic. Therefore the finality considered here is that
the user can understand the different sentences transcribed and synthesized which will be
pronounced with a clear and high quality way.
2. OVERVIEW OF SPEECH SYNTHESIS SYSTEMS
At the present time, we can judge that works carried out in the same context of ours are still not
really colossal, and this is because of the complexity of the language itself. And if this some
works exists, they are based on the same principle as transcribers of the other languages
(French, English, etc.) [1], nevertheless, the efforts undertaken are encouraging and open a large
window to follow research task in this field.
International Journal on Natural Language Computing (IJNLC) Vol. 3, No.2, April 2014
2
− TTS system MBROLA [2] which use the code SAMPA during the stage of
transcription, in this case the user must respect the form of the SAMPA code which
is not a universal code;
− Work of S. BALOUL [3] who represents a concrete example of transcription of words;
based on morphological analysis, and on the studies of pauses to generate the
pronunciation of the texts.
− SYAMSA (SYstème d’Analyse Morphosyntaxique de l’Arabe), realized by SAROH
[4]. According to him, "the phonetisation of the Arabic language is based in particularly
on the use of lexicons and on a morphological analyzer for the generation of the
different forms of a word. In addition, they are the phenomena of interaction among the
words (connection, elision, etc.) and the phenomena of assimilation which suggest the
uses of phonological rules "[5]. This tool ensures for each word in entry, the root which
correspond to it as well as the morphological and phonetic representations of the root.
− The GHAZALI [6] project which was carried out within the IRSIT (Institut Régional
des Sciences Informatiques et des Télécommunications) of Tunisia, it is based on a
transcription work which fits inside the framework of the realization of a TTS system.
The characteristic of this system is shown in the use of a set of rules in the emphasis
propagation.
− SYNTHAR+ [7] which is a tool studied by Z. ZEMIRLI within the NII (the National
Institute of Informatics of Algiers), it ensures the transcription step for a TTS system so
that it transmits the phonetic representation to the MULTIVOX synthesizer. It should be
known that SYNTHAR+ is based on a morphological analysis before realizing the
transcription.
3. DESIGN OF THE SYSTEM
The uses of the concept of code and the introduction of high levels in analysis (morphological,
syntactic, and pragmatic.) makes the transcription task so difficult and requires deepened studies
of the language itself. The difference in our work compared to all that exists is in the
transcription using graphemes, i.e. the uses of the Arabic characters as basic units to transcribe
directly the text, indeed modeled by an expert system.
Thus the role of an expert system is then to infer generally rules like if/then. To build this
system our work targets the following steps:
International Journal on Natural Language Computing (IJNLC) Vol. 3, No.2, April 2014
3
Figure 1. General schema of our system of speech synthesis
3.1. Acquisition
It is the first step of our process of voice synthesis; it plays the role of the interface intermediate
between the user and the system. During this time the user have to enter his set of phrases to be
pronounced by the system. After that the automatic process can begin.
3.2. Pre-processing and standardization of the text
In this phase any form of text is transformed to it literal, perfectly disambiguated form [1]. In
particular, this module draft problems of formats, dates and hours, abbreviations, numbers,
currencies, addresses, emails…etc. It was shown that majority errors of conversion
grapheme/phoneme, for the best operational systems came from the proper names and the
exceptions that they pose [8]. Here are some examples of the processing carried out in our case:
− The replacement of each composed character by its equivalents
Example: ‫ا‬ ‫ل‬
− Consultation of the exceptions lexicon to eliminate the special words.
− The application of the transcription rules established for the language. This module
must be able to process grammatical complexes tasks to identify the rules of
transcription which will be used in a considered context.
3.3. Transcription
Two approaches are generally used here:
− The use of the lexicon:
In this case we must assign to each word in input, the pronunciation which corresponds
to it without taking in account its context. The speed, flexibility and simplicity are the
main advantages of this approach.
It has been shown that the majority errors of conversion grapheme/phoneme, for the
best operating systems came from the proper names and the exceptions that they
International Journal on Natural Language Computing (IJNLC) Vol. 3, No.2, April 2014
4
pose [9]. Exceptions are words that are not read based on well determined rules of
writing; because of this they represent the famous example of this approach.
Table 1: a few words of exceptions [5]
− The use of rules
In this approach each grapheme is converted to a phoneme depending on the context
using a set of rewriting rules. The main advantage of this approach come out in the
ability to model the linguistic knowledge of human beings by a set of rules that can be
incorporated in expert systems. Each of these rules has the following form:
[Phoneme] = {LG (LeftContext)} + {C (character)} + {RD (Right Context)}
Here is a concrete example of transcription rules [11]:
# Is a sign of beginning of sentence,
$ Is a sign of end of sentence,
§ Is one end of word,
C Is a consonant,
V Is a vowel,
SC Is a Solar Consonant and LC is a Lunar Consonant
[UU] = {CS} + {‫و‬} +{ ُ} [uu]= §+ {} + { ُ‫و‬ }
[UU] = {CL} + {‫و‬} +{ ُ} [uu]= $ {‫و‬ } + { ُ }
- -When the " ‫و‬ " is preceded by the vowel/ ُ/ and followed by a consonant, we
obtain the long vowel [uu].
- When the " ‫و‬ " is preceded by the vowel / ُ/ in final position, we obtain the long
vowel [uu].
The second type represents the diphones which are extracted from the set of artificial words and
which do not have a meaning, also called logatomes, or words porters. Each of these is used to
extract a single diphone i.e. each logatome contains one and only one diphone and this in order
to ensure a degree of independence between this diphone and its context. Here are a few
examples of logatomes used :
C represents a consonant
V a vowel
# A silence for the beginning or end of word
International Journal on Natural Language Computing (IJNLC) Vol. 3, No.2, April 2014
5
3.4. Analysis (the extraction of the characteristics)
The aim of the analysis of the voice signal is to extracts the acoustic vectors which will
be used in the stage of synthesis follows. In this step the voice signal is transformed into a
sequence of acoustic vectors on the way to decrease redundancy and the amount of data to be
processed. And then a spectral analysis by the discrete Fourier transform is performed on a
signal frame (typically of size 20 or 30 ms) [9]. In this frame, the voice signal is considered to
be sufficiently stable and we extract a vector of parameters considered to be enough for the
good operation of the voice signal. In the speech synthesis, the characteristics extraction step,
commonly known as the step of analysis, can be achieved in several ways. Indeed, the acoustic
vectors are usually extracted using methods such as temporal encoding predictive linear (Linear
Predictive Coding LPC) or Cepstrales methods as the MFCC encoding (Mel Frequency Cepstral
Coding), as well as the process of segmentation, etc. This process delimits on the acoustic signal
a set of segments characterized by labels belonging to the phonetics alphabet of the language
under consideration.
At present, the segmentation completely automatic of a voice signal remains a fastidious task.
Indeed, looking at the complexity acoustico-phonetics phenomena being studied, this activity
requires often a manual intervention. Generally, the methods that perform the segmentation of
the acoustic wave are divided into two great classes:
− The first include all the methods which allow segmenting a voice signal without a priori
knowledge of the linguistics content of this signal. These methods split the voice signal
into a set of zones homogeneous and stable;
− The second class includes all the methods which allow segmenting the voice signal
basing on a priori linguistic description (typically on phonemes) of this signal. These
methods of segmentation are revealed like methods with linguistically constraints. In
our state, we have opted for the second technique using a tool of voice signal analysis
which is PRAAT [10] so to slice manually the speech signal in a succession of
segments, each one associated with an element acoustic unit (phoneme or diaphone).
International Journal on Natural Language Computing (IJNLC) Vol. 3, No.2, April 2014
6
Figure 2. Decomposition in polysound of the word [ ]
3.5. Creation of the sound database
Majority of work curried out in the field of the spoken communication required often the
recording, and the handling of corpuses of continuous speech, and that to curry out studies on
the contextual effects, one the phonetic indices, and variability intra and inter-speaker. Our
corpus is modeled by the diagram of class1 as follow:
Figure 3. The diagram of class of our sound database
3.6. Voice generation
This step commonly known as speech synthesis consist of a transformation of the phonetic
sequence (which represents the pronunciation of the written text) resulting from the
1: It is a diagram used in UML(Unified Modeling Language) which is used in object oriented modeling
International Journal on Natural Language Computing (IJNLC) Vol. 3, No.2, April 2014
7
transcription step to its substance, i.e. to it acoustics realization. During this step we choose
from our sound database the units (phonemes, diphones) most suitable to build the sentence to
generate, this means that we will create a automatic function of reading, so that in the end, the
user has only to listen to the synthetic sentences. To improve the quality of the generated
synthesis, we increase each time the sound unit only. The general silhouette of our expert
system can be as follow:
Figure 4. General diagram of our expert system
In our case we have used the method of synthesis by concatenation and here is our reading
function:
position= find (grapheme[ ig],API ) ;
if((grapheme[ig]==‘‫&&)'أ‬ (grapheme[ig+1]==‘‫))'ل‬
{
mp2->FileName="C:son_hananealif.wav";
mp2->Open();
mp2->Wait=true;
mp2->Play();
ig=ig+2;
position=find (grapheme[ig],API);
if(API[position][1]=='L')
{
mp2->FileName="C:son_hananel.wav";
mp2->Open();
mp2->Wait=true;
mp2->Play();
mp2->FileName=API[position][2];
mp2->Open();
mp2->Wait=true;
mp2->Play();
ig++;
}
International Journal on Natural Language Computing (IJNLC) Vol. 3, No.2, April 2014
8
else
{
if(API[position][1]=='S')
{
mp2->FileName=API[position][2];
mp2->Open();
mp2->Wait=true;
mp2->Play();
ig++;
}
}
}
4. TESTS AND RESULTS
To test the performances of our TTS system based on Standard Arabic language, we have
chosen a set of sentences which we judged like reference since they contain almost the different
possible combinations specific to the language itself. To calculate the success rate (SR)
associated with each sentence tested; we got the following formula:
%100*
drasestesteNumberofph
ononcedrasewellprNumberofph
SR =
The system present in general a SR of 96 % for the set of the sentences tested. Results obtained
are summarized by the following table:
Table 2. Rate of success for a sample of selected sentences
Majority
content
POT Synthesis by
Phonemes
Synthesis by
Diphones
Short vowels 100% 95% /
Long vowels 100% 95% /
Solar consonants 100% 97% /
Lunar
consonants
100% 95% /
Isolated words 100% 80% 90%
Sentences 100% 75% 85%
Numbers 90% 95% 100%
Exception Words 100% / /
5. CONCLUSION
Synthesis systems from text written in Standard Arabic represent a research area very active but
difficult to implement. In order to facilitate this task we have proposed a tool, an expert system
for voice generation based on Standard Arabic text has been developed, this system uses a set of
sub phonetic elements as the synthesis units to allow synthesis of limited-vocabulary speech of
good quality. The synthesis units have been defined after a careful study of the phonetic
International Journal on Natural Language Computing (IJNLC) Vol. 3, No.2, April 2014
9
properties of modern Standard Arabic, and they consist of central steady-state portions of
vowels, central steady-state portions of consonants, vowel-consonant and consonant-vowel
transitions, and some allophones. A text-to-speech system which uses this method has also been
explored. The input of the system is usual Arabic spelling with diacritics and/or simple numeric
expressions.
Synthesis is controlled by several text-to-speech rules within the rule database of the expert
system, which are formulated and developed as algorithms more suited for computer handling
of the synthesis process. The rules are required for converting the input text into phonemes,
converting the phonemes into phonetic units, generating the synthesis units from the phonetic
units, and concatenating the synthesis units to form spoken messages. The suitability of the
method for synthesizing Arabic has been shown by realizing all its functions on a personal
computer and by conducting understandability test on synthesized speech. So, we have detailed
the different components which represents the basic blocks of our TTS system based on a
written text in Standard Arabic and our modeling of it with the use of an expert system, this tool
which is a fruits of the artificial intelligence remain less used in the field of automatic speech
processing. This fact encouraged us to explore this world trying to give a little push to the
research done in this multidisciplinary field.
REFERENCES
[1] Tebbi Hanane, « Transcription orthographique phonétique en vue de la synthèse de la parole a
partir du texte de l’Arabe », Mémoire de magister en Informatique, Blida, Algérie, 2007
[2] http//:tcts.fpms.ac.be/ synthesis/ Mbrola.html.
[3] Baloul, « Développement d’un système automatique de synthèse de la parole à partir du texte
arabe standard voyellé ». Thèse de doctorat d’université, Le Mans, France, 27 Mai 2003.
[4] SAROH Abdelghani, Base de données lexicales dans un système d'analyse morpho-syntaxique
de l'arabe : SYAMSA, Toulouse 3, 1989.
[5] Guerti Mhania, Tebbi Hanane, La Conversion Graphèmes Phonèmes En Vue D'une Lecture
Automatique de Textes en Arabe Standard, LANIA, Chlef, Algérie, 2007 S.
[6] Dichy J. & Hassoun M.O. (1998), Some aspects of the DIINAR-MBC research programme, in
Ubaydly A., ed.,1998: 2.8.1-5.
[7] SYNTHAR+: Arabic speech synthesis under multivox. (SYNTHAR+: Synthèse vocale arabe
sous Multivox.) (English) RAIRO, Tech. Sci. Inf. 17, No. 6, 741-761 (1998).
[8] Mehdi Braham, Oumaya Abbes, Maroua Trabelsi, Mehdi Dahmen, Rania Nouaari, « Intelligence
artificielle : Diagnostic par système expert, Modèle d’analyse d’acteurs : MACTOR »
[9] P. Boula de Mareüil, « Synthèse de la parole à partir de courriers et évaluation de la conversion
graphème-phonème ». LIMSI-CNRS http://www.limsi.fr/ Individu/ mareuil/
[10] Paul Boersma and David Weenink, "Praat: doing phonetics by computer" Phonetic Sciences,
University of Amsterdam Spuistraat 210, 1012VT Amsterdam The Netherlands.
[11] T. Saidane, M. Zrigui, & M. Ben Ahmed, « La transcription orthographique phonétique de la
langue Arabe ». RECITAL 2004, Fès, 19-22 Avril 2004.

Contenu connexe

Tendances

Error Detection and Feedback with OT-LFG for Computer-assisted Language Learning
Error Detection and Feedback with OT-LFG for Computer-assisted Language LearningError Detection and Feedback with OT-LFG for Computer-assisted Language Learning
Error Detection and Feedback with OT-LFG for Computer-assisted Language LearningCITE
 
Comparison of Feature Extraction MFCC and LPC in Automatic Speech Recognition...
Comparison of Feature Extraction MFCC and LPC in Automatic Speech Recognition...Comparison of Feature Extraction MFCC and LPC in Automatic Speech Recognition...
Comparison of Feature Extraction MFCC and LPC in Automatic Speech Recognition...TELKOMNIKA JOURNAL
 
A Marathi Hidden-Markov Model Based Speech Synthesis System
A Marathi Hidden-Markov Model Based Speech Synthesis SystemA Marathi Hidden-Markov Model Based Speech Synthesis System
A Marathi Hidden-Markov Model Based Speech Synthesis Systemiosrjce
 
33 9765 development paper id 0034 (edit a) (1)
33 9765 development paper id 0034 (edit a) (1)33 9765 development paper id 0034 (edit a) (1)
33 9765 development paper id 0034 (edit a) (1)IAESIJEECS
 
Wavelet Based Feature Extraction for the Indonesian CV Syllables Sound
Wavelet Based Feature Extraction for the Indonesian CV Syllables SoundWavelet Based Feature Extraction for the Indonesian CV Syllables Sound
Wavelet Based Feature Extraction for the Indonesian CV Syllables SoundTELKOMNIKA JOURNAL
 
HINDI AND MARATHI TO ENGLISH MACHINE TRANSLITERATION USING SVM
HINDI AND MARATHI TO ENGLISH MACHINE TRANSLITERATION USING SVMHINDI AND MARATHI TO ENGLISH MACHINE TRANSLITERATION USING SVM
HINDI AND MARATHI TO ENGLISH MACHINE TRANSLITERATION USING SVMijnlc
 
Classification improvement of spoken arabic language based on radial basis fu...
Classification improvement of spoken arabic language based on radial basis fu...Classification improvement of spoken arabic language based on radial basis fu...
Classification improvement of spoken arabic language based on radial basis fu...IJECEIAES
 
International journal of signal and image processing issues vol 2015 - no 1...
International journal of signal and image processing issues   vol 2015 - no 1...International journal of signal and image processing issues   vol 2015 - no 1...
International journal of signal and image processing issues vol 2015 - no 1...sophiabelthome
 
SPEAKER VERIFICATION USING ACOUSTIC AND PROSODIC FEATURES
SPEAKER VERIFICATION USING ACOUSTIC AND PROSODIC FEATURESSPEAKER VERIFICATION USING ACOUSTIC AND PROSODIC FEATURES
SPEAKER VERIFICATION USING ACOUSTIC AND PROSODIC FEATURESacijjournal
 
Artificially Generatedof Concatenative Syllable based Text to Speech Synthesi...
Artificially Generatedof Concatenative Syllable based Text to Speech Synthesi...Artificially Generatedof Concatenative Syllable based Text to Speech Synthesi...
Artificially Generatedof Concatenative Syllable based Text to Speech Synthesi...iosrjce
 
Speech to text conversion for visually impaired person using µ law companding
Speech to text conversion for visually impaired person using µ law compandingSpeech to text conversion for visually impaired person using µ law companding
Speech to text conversion for visually impaired person using µ law compandingiosrjce
 
Deep Learning勉強会@小町研 "Learning Character-level Representations for Part-of-Sp...
Deep Learning勉強会@小町研 "Learning Character-level Representations for Part-of-Sp...Deep Learning勉強会@小町研 "Learning Character-level Representations for Part-of-Sp...
Deep Learning勉強会@小町研 "Learning Character-level Representations for Part-of-Sp...Yuki Tomo
 
Isolated word recognition using lpc & vector quantization
Isolated word recognition using lpc & vector quantizationIsolated word recognition using lpc & vector quantization
Isolated word recognition using lpc & vector quantizationeSAT Publishing House
 
IMPROVING MYANMAR AUTOMATIC SPEECH RECOGNITION WITH OPTIMIZATION OF CONVOLUTI...
IMPROVING MYANMAR AUTOMATIC SPEECH RECOGNITION WITH OPTIMIZATION OF CONVOLUTI...IMPROVING MYANMAR AUTOMATIC SPEECH RECOGNITION WITH OPTIMIZATION OF CONVOLUTI...
IMPROVING MYANMAR AUTOMATIC SPEECH RECOGNITION WITH OPTIMIZATION OF CONVOLUTI...ijnlc
 

Tendances (15)

[IJET-V2I1P13] Authors:Shilpa More, Gagandeep .S. Dhir , Deepak Daiwadney and...
[IJET-V2I1P13] Authors:Shilpa More, Gagandeep .S. Dhir , Deepak Daiwadney and...[IJET-V2I1P13] Authors:Shilpa More, Gagandeep .S. Dhir , Deepak Daiwadney and...
[IJET-V2I1P13] Authors:Shilpa More, Gagandeep .S. Dhir , Deepak Daiwadney and...
 
Error Detection and Feedback with OT-LFG for Computer-assisted Language Learning
Error Detection and Feedback with OT-LFG for Computer-assisted Language LearningError Detection and Feedback with OT-LFG for Computer-assisted Language Learning
Error Detection and Feedback with OT-LFG for Computer-assisted Language Learning
 
Comparison of Feature Extraction MFCC and LPC in Automatic Speech Recognition...
Comparison of Feature Extraction MFCC and LPC in Automatic Speech Recognition...Comparison of Feature Extraction MFCC and LPC in Automatic Speech Recognition...
Comparison of Feature Extraction MFCC and LPC in Automatic Speech Recognition...
 
A Marathi Hidden-Markov Model Based Speech Synthesis System
A Marathi Hidden-Markov Model Based Speech Synthesis SystemA Marathi Hidden-Markov Model Based Speech Synthesis System
A Marathi Hidden-Markov Model Based Speech Synthesis System
 
33 9765 development paper id 0034 (edit a) (1)
33 9765 development paper id 0034 (edit a) (1)33 9765 development paper id 0034 (edit a) (1)
33 9765 development paper id 0034 (edit a) (1)
 
Wavelet Based Feature Extraction for the Indonesian CV Syllables Sound
Wavelet Based Feature Extraction for the Indonesian CV Syllables SoundWavelet Based Feature Extraction for the Indonesian CV Syllables Sound
Wavelet Based Feature Extraction for the Indonesian CV Syllables Sound
 
HINDI AND MARATHI TO ENGLISH MACHINE TRANSLITERATION USING SVM
HINDI AND MARATHI TO ENGLISH MACHINE TRANSLITERATION USING SVMHINDI AND MARATHI TO ENGLISH MACHINE TRANSLITERATION USING SVM
HINDI AND MARATHI TO ENGLISH MACHINE TRANSLITERATION USING SVM
 
Classification improvement of spoken arabic language based on radial basis fu...
Classification improvement of spoken arabic language based on radial basis fu...Classification improvement of spoken arabic language based on radial basis fu...
Classification improvement of spoken arabic language based on radial basis fu...
 
International journal of signal and image processing issues vol 2015 - no 1...
International journal of signal and image processing issues   vol 2015 - no 1...International journal of signal and image processing issues   vol 2015 - no 1...
International journal of signal and image processing issues vol 2015 - no 1...
 
SPEAKER VERIFICATION USING ACOUSTIC AND PROSODIC FEATURES
SPEAKER VERIFICATION USING ACOUSTIC AND PROSODIC FEATURESSPEAKER VERIFICATION USING ACOUSTIC AND PROSODIC FEATURES
SPEAKER VERIFICATION USING ACOUSTIC AND PROSODIC FEATURES
 
Artificially Generatedof Concatenative Syllable based Text to Speech Synthesi...
Artificially Generatedof Concatenative Syllable based Text to Speech Synthesi...Artificially Generatedof Concatenative Syllable based Text to Speech Synthesi...
Artificially Generatedof Concatenative Syllable based Text to Speech Synthesi...
 
Speech to text conversion for visually impaired person using µ law companding
Speech to text conversion for visually impaired person using µ law compandingSpeech to text conversion for visually impaired person using µ law companding
Speech to text conversion for visually impaired person using µ law companding
 
Deep Learning勉強会@小町研 "Learning Character-level Representations for Part-of-Sp...
Deep Learning勉強会@小町研 "Learning Character-level Representations for Part-of-Sp...Deep Learning勉強会@小町研 "Learning Character-level Representations for Part-of-Sp...
Deep Learning勉強会@小町研 "Learning Character-level Representations for Part-of-Sp...
 
Isolated word recognition using lpc & vector quantization
Isolated word recognition using lpc & vector quantizationIsolated word recognition using lpc & vector quantization
Isolated word recognition using lpc & vector quantization
 
IMPROVING MYANMAR AUTOMATIC SPEECH RECOGNITION WITH OPTIMIZATION OF CONVOLUTI...
IMPROVING MYANMAR AUTOMATIC SPEECH RECOGNITION WITH OPTIMIZATION OF CONVOLUTI...IMPROVING MYANMAR AUTOMATIC SPEECH RECOGNITION WITH OPTIMIZATION OF CONVOLUTI...
IMPROVING MYANMAR AUTOMATIC SPEECH RECOGNITION WITH OPTIMIZATION OF CONVOLUTI...
 

En vedette

Building a vietnamese dialog mechanism for v dlg~tabl system
Building a vietnamese dialog mechanism for v dlg~tabl systemBuilding a vietnamese dialog mechanism for v dlg~tabl system
Building a vietnamese dialog mechanism for v dlg~tabl systemijnlc
 
Verb based manipuri sentiment analysis
Verb based manipuri sentiment analysisVerb based manipuri sentiment analysis
Verb based manipuri sentiment analysisijnlc
 
Smart grammar a dynamic spoken language understanding grammar for inflective ...
Smart grammar a dynamic spoken language understanding grammar for inflective ...Smart grammar a dynamic spoken language understanding grammar for inflective ...
Smart grammar a dynamic spoken language understanding grammar for inflective ...ijnlc
 
A MULTI-STREAM HMM APPROACH TO OFFLINE HANDWRITTEN ARABIC WORD RECOGNITION
A MULTI-STREAM HMM APPROACH TO OFFLINE HANDWRITTEN ARABIC WORD RECOGNITIONA MULTI-STREAM HMM APPROACH TO OFFLINE HANDWRITTEN ARABIC WORD RECOGNITION
A MULTI-STREAM HMM APPROACH TO OFFLINE HANDWRITTEN ARABIC WORD RECOGNITIONijnlc
 
CLUSTERING WEB SEARCH RESULTS FOR EFFECTIVE ARABIC LANGUAGE BROWSING
CLUSTERING WEB SEARCH RESULTS FOR EFFECTIVE ARABIC LANGUAGE BROWSINGCLUSTERING WEB SEARCH RESULTS FOR EFFECTIVE ARABIC LANGUAGE BROWSING
CLUSTERING WEB SEARCH RESULTS FOR EFFECTIVE ARABIC LANGUAGE BROWSINGijnlc
 
HANDLING UNKNOWN WORDS IN NAMED ENTITY RECOGNITION USING TRANSLITERATION
HANDLING UNKNOWN WORDS IN NAMED ENTITY RECOGNITION USING TRANSLITERATIONHANDLING UNKNOWN WORDS IN NAMED ENTITY RECOGNITION USING TRANSLITERATION
HANDLING UNKNOWN WORDS IN NAMED ENTITY RECOGNITION USING TRANSLITERATIONijnlc
 
Hybrid part of-speech tagger for non-vocalized arabic text
Hybrid part of-speech tagger for non-vocalized arabic textHybrid part of-speech tagger for non-vocalized arabic text
Hybrid part of-speech tagger for non-vocalized arabic textijnlc
 
IMPLEMENTATION OF NLIZATION FRAMEWORK FOR VERBS, PRONOUNS AND DETERMINERS WIT...
IMPLEMENTATION OF NLIZATION FRAMEWORK FOR VERBS, PRONOUNS AND DETERMINERS WIT...IMPLEMENTATION OF NLIZATION FRAMEWORK FOR VERBS, PRONOUNS AND DETERMINERS WIT...
IMPLEMENTATION OF NLIZATION FRAMEWORK FOR VERBS, PRONOUNS AND DETERMINERS WIT...ijnlc
 
Chunking in manipuri using crf
Chunking in manipuri using crfChunking in manipuri using crf
Chunking in manipuri using crfijnlc
 
Developemnt and evaluation of a web based question answering system for arabi...
Developemnt and evaluation of a web based question answering system for arabi...Developemnt and evaluation of a web based question answering system for arabi...
Developemnt and evaluation of a web based question answering system for arabi...ijnlc
 
NERHMM: A TOOL FOR NAMED ENTITY RECOGNITION BASED ON HIDDEN MARKOV MODEL
NERHMM: A TOOL FOR NAMED ENTITY RECOGNITION BASED ON HIDDEN MARKOV MODELNERHMM: A TOOL FOR NAMED ENTITY RECOGNITION BASED ON HIDDEN MARKOV MODEL
NERHMM: A TOOL FOR NAMED ENTITY RECOGNITION BASED ON HIDDEN MARKOV MODELijnlc
 
An improved apriori algorithm for association rules
An improved apriori algorithm for association rulesAn improved apriori algorithm for association rules
An improved apriori algorithm for association rulesijnlc
 
A comparative analysis of particle swarm optimization and k means algorithm f...
A comparative analysis of particle swarm optimization and k means algorithm f...A comparative analysis of particle swarm optimization and k means algorithm f...
A comparative analysis of particle swarm optimization and k means algorithm f...ijnlc
 
IMPROVING THE QUALITY OF GUJARATI-HINDI MACHINE TRANSLATION THROUGH PART-OF-S...
IMPROVING THE QUALITY OF GUJARATI-HINDI MACHINE TRANSLATION THROUGH PART-OF-S...IMPROVING THE QUALITY OF GUJARATI-HINDI MACHINE TRANSLATION THROUGH PART-OF-S...
IMPROVING THE QUALITY OF GUJARATI-HINDI MACHINE TRANSLATION THROUGH PART-OF-S...ijnlc
 
An exhaustive font and size invariant classification scheme for ocr of devana...
An exhaustive font and size invariant classification scheme for ocr of devana...An exhaustive font and size invariant classification scheme for ocr of devana...
An exhaustive font and size invariant classification scheme for ocr of devana...ijnlc
 
Evaluation of subjective answers using glsa enhanced with contextual synonymy
Evaluation of subjective answers using glsa enhanced with contextual synonymyEvaluation of subjective answers using glsa enhanced with contextual synonymy
Evaluation of subjective answers using glsa enhanced with contextual synonymyijnlc
 
S ENTIMENT A NALYSIS F OR M ODERN S TANDARD A RABIC A ND C OLLOQUIAl
S ENTIMENT A NALYSIS  F OR M ODERN S TANDARD  A RABIC  A ND  C OLLOQUIAlS ENTIMENT A NALYSIS  F OR M ODERN S TANDARD  A RABIC  A ND  C OLLOQUIAl
S ENTIMENT A NALYSIS F OR M ODERN S TANDARD A RABIC A ND C OLLOQUIAlijnlc
 
S URVEY O N M ACHINE T RANSLITERATION A ND M ACHINE L EARNING M ODELS
S URVEY  O N M ACHINE  T RANSLITERATION A ND  M ACHINE L EARNING M ODELSS URVEY  O N M ACHINE  T RANSLITERATION A ND  M ACHINE L EARNING M ODELS
S URVEY O N M ACHINE T RANSLITERATION A ND M ACHINE L EARNING M ODELSijnlc
 
Conceptual framework for abstractive text summarization
Conceptual framework for abstractive text summarizationConceptual framework for abstractive text summarization
Conceptual framework for abstractive text summarizationijnlc
 

En vedette (20)

Building a vietnamese dialog mechanism for v dlg~tabl system
Building a vietnamese dialog mechanism for v dlg~tabl systemBuilding a vietnamese dialog mechanism for v dlg~tabl system
Building a vietnamese dialog mechanism for v dlg~tabl system
 
Verb based manipuri sentiment analysis
Verb based manipuri sentiment analysisVerb based manipuri sentiment analysis
Verb based manipuri sentiment analysis
 
Smart grammar a dynamic spoken language understanding grammar for inflective ...
Smart grammar a dynamic spoken language understanding grammar for inflective ...Smart grammar a dynamic spoken language understanding grammar for inflective ...
Smart grammar a dynamic spoken language understanding grammar for inflective ...
 
A MULTI-STREAM HMM APPROACH TO OFFLINE HANDWRITTEN ARABIC WORD RECOGNITION
A MULTI-STREAM HMM APPROACH TO OFFLINE HANDWRITTEN ARABIC WORD RECOGNITIONA MULTI-STREAM HMM APPROACH TO OFFLINE HANDWRITTEN ARABIC WORD RECOGNITION
A MULTI-STREAM HMM APPROACH TO OFFLINE HANDWRITTEN ARABIC WORD RECOGNITION
 
CLUSTERING WEB SEARCH RESULTS FOR EFFECTIVE ARABIC LANGUAGE BROWSING
CLUSTERING WEB SEARCH RESULTS FOR EFFECTIVE ARABIC LANGUAGE BROWSINGCLUSTERING WEB SEARCH RESULTS FOR EFFECTIVE ARABIC LANGUAGE BROWSING
CLUSTERING WEB SEARCH RESULTS FOR EFFECTIVE ARABIC LANGUAGE BROWSING
 
HANDLING UNKNOWN WORDS IN NAMED ENTITY RECOGNITION USING TRANSLITERATION
HANDLING UNKNOWN WORDS IN NAMED ENTITY RECOGNITION USING TRANSLITERATIONHANDLING UNKNOWN WORDS IN NAMED ENTITY RECOGNITION USING TRANSLITERATION
HANDLING UNKNOWN WORDS IN NAMED ENTITY RECOGNITION USING TRANSLITERATION
 
Hybrid part of-speech tagger for non-vocalized arabic text
Hybrid part of-speech tagger for non-vocalized arabic textHybrid part of-speech tagger for non-vocalized arabic text
Hybrid part of-speech tagger for non-vocalized arabic text
 
IMPLEMENTATION OF NLIZATION FRAMEWORK FOR VERBS, PRONOUNS AND DETERMINERS WIT...
IMPLEMENTATION OF NLIZATION FRAMEWORK FOR VERBS, PRONOUNS AND DETERMINERS WIT...IMPLEMENTATION OF NLIZATION FRAMEWORK FOR VERBS, PRONOUNS AND DETERMINERS WIT...
IMPLEMENTATION OF NLIZATION FRAMEWORK FOR VERBS, PRONOUNS AND DETERMINERS WIT...
 
Chunking in manipuri using crf
Chunking in manipuri using crfChunking in manipuri using crf
Chunking in manipuri using crf
 
Developemnt and evaluation of a web based question answering system for arabi...
Developemnt and evaluation of a web based question answering system for arabi...Developemnt and evaluation of a web based question answering system for arabi...
Developemnt and evaluation of a web based question answering system for arabi...
 
NERHMM: A TOOL FOR NAMED ENTITY RECOGNITION BASED ON HIDDEN MARKOV MODEL
NERHMM: A TOOL FOR NAMED ENTITY RECOGNITION BASED ON HIDDEN MARKOV MODELNERHMM: A TOOL FOR NAMED ENTITY RECOGNITION BASED ON HIDDEN MARKOV MODEL
NERHMM: A TOOL FOR NAMED ENTITY RECOGNITION BASED ON HIDDEN MARKOV MODEL
 
An improved apriori algorithm for association rules
An improved apriori algorithm for association rulesAn improved apriori algorithm for association rules
An improved apriori algorithm for association rules
 
A comparative analysis of particle swarm optimization and k means algorithm f...
A comparative analysis of particle swarm optimization and k means algorithm f...A comparative analysis of particle swarm optimization and k means algorithm f...
A comparative analysis of particle swarm optimization and k means algorithm f...
 
IMPROVING THE QUALITY OF GUJARATI-HINDI MACHINE TRANSLATION THROUGH PART-OF-S...
IMPROVING THE QUALITY OF GUJARATI-HINDI MACHINE TRANSLATION THROUGH PART-OF-S...IMPROVING THE QUALITY OF GUJARATI-HINDI MACHINE TRANSLATION THROUGH PART-OF-S...
IMPROVING THE QUALITY OF GUJARATI-HINDI MACHINE TRANSLATION THROUGH PART-OF-S...
 
An exhaustive font and size invariant classification scheme for ocr of devana...
An exhaustive font and size invariant classification scheme for ocr of devana...An exhaustive font and size invariant classification scheme for ocr of devana...
An exhaustive font and size invariant classification scheme for ocr of devana...
 
Evaluation of subjective answers using glsa enhanced with contextual synonymy
Evaluation of subjective answers using glsa enhanced with contextual synonymyEvaluation of subjective answers using glsa enhanced with contextual synonymy
Evaluation of subjective answers using glsa enhanced with contextual synonymy
 
S ENTIMENT A NALYSIS F OR M ODERN S TANDARD A RABIC A ND C OLLOQUIAl
S ENTIMENT A NALYSIS  F OR M ODERN S TANDARD  A RABIC  A ND  C OLLOQUIAlS ENTIMENT A NALYSIS  F OR M ODERN S TANDARD  A RABIC  A ND  C OLLOQUIAl
S ENTIMENT A NALYSIS F OR M ODERN S TANDARD A RABIC A ND C OLLOQUIAl
 
S URVEY O N M ACHINE T RANSLITERATION A ND M ACHINE L EARNING M ODELS
S URVEY  O N M ACHINE  T RANSLITERATION A ND  M ACHINE L EARNING M ODELSS URVEY  O N M ACHINE  T RANSLITERATION A ND  M ACHINE L EARNING M ODELS
S URVEY O N M ACHINE T RANSLITERATION A ND M ACHINE L EARNING M ODELS
 
Conceptual framework for abstractive text summarization
Conceptual framework for abstractive text summarizationConceptual framework for abstractive text summarization
Conceptual framework for abstractive text summarization
 
Document excel
Document excelDocument excel
Document excel
 

Similaire à An expert system for automatic reading of a text written in standard arabic

SYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMAR
SYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMARSYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMAR
SYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMARijcseit
 
SYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMAR
SYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMARSYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMAR
SYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMARijcseit
 
SYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMAR
SYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMARSYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMAR
SYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMARijcseit
 
SYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMAR
SYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMARSYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMAR
SYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMARijcseit
 
High Quality Arabic Concatenative Speech Synthesis
High Quality Arabic Concatenative Speech SynthesisHigh Quality Arabic Concatenative Speech Synthesis
High Quality Arabic Concatenative Speech Synthesissipij
 
Survey On Speech Synthesis
Survey On Speech SynthesisSurvey On Speech Synthesis
Survey On Speech SynthesisCSCJournals
 
Emotional telugu speech signals classification based on k nn classifier
Emotional telugu speech signals classification based on k nn classifierEmotional telugu speech signals classification based on k nn classifier
Emotional telugu speech signals classification based on k nn classifiereSAT Publishing House
 
Emotional telugu speech signals classification based on k nn classifier
Emotional telugu speech signals classification based on k nn classifierEmotional telugu speech signals classification based on k nn classifier
Emotional telugu speech signals classification based on k nn classifiereSAT Journals
 
Direct Punjabi to English Speech Translation using Discrete Units
Direct Punjabi to English Speech Translation using Discrete UnitsDirect Punjabi to English Speech Translation using Discrete Units
Direct Punjabi to English Speech Translation using Discrete UnitsIJCI JOURNAL
 
PUNJABI SPEECH SYNTHESIS SYSTEM USING HTK
PUNJABI SPEECH SYNTHESIS SYSTEM USING HTKPUNJABI SPEECH SYNTHESIS SYSTEM USING HTK
PUNJABI SPEECH SYNTHESIS SYSTEM USING HTKijistjournal
 
Development of text to speech system for yoruba language
Development of text to speech system for yoruba languageDevelopment of text to speech system for yoruba language
Development of text to speech system for yoruba languageAlexander Decker
 
EFFECT OF DYNAMIC TIME WARPING ON ALIGNMENT OF PHRASES AND PHONEMES
EFFECT OF DYNAMIC TIME WARPING ON ALIGNMENT OF PHRASES AND PHONEMESEFFECT OF DYNAMIC TIME WARPING ON ALIGNMENT OF PHRASES AND PHONEMES
EFFECT OF DYNAMIC TIME WARPING ON ALIGNMENT OF PHRASES AND PHONEMESkevig
 
Tutorial - Speech Synthesis System
Tutorial - Speech Synthesis SystemTutorial - Speech Synthesis System
Tutorial - Speech Synthesis SystemIJERA Editor
 
Effect of Dynamic Time Warping on Alignment of Phrases and Phonemes
Effect of Dynamic Time Warping on Alignment of Phrases and PhonemesEffect of Dynamic Time Warping on Alignment of Phrases and Phonemes
Effect of Dynamic Time Warping on Alignment of Phrases and Phonemeskevig
 
E0502 01 2327
E0502 01 2327E0502 01 2327
E0502 01 2327IJMER
 
Acceptance Testing Of A Spoken Language Translation System
Acceptance Testing Of A Spoken Language Translation SystemAcceptance Testing Of A Spoken Language Translation System
Acceptance Testing Of A Spoken Language Translation SystemMichele Thomas
 
Implementation of English-Text to Marathi-Speech (ETMS) Synthesizer
Implementation of English-Text to Marathi-Speech (ETMS) SynthesizerImplementation of English-Text to Marathi-Speech (ETMS) Synthesizer
Implementation of English-Text to Marathi-Speech (ETMS) SynthesizerIOSR Journals
 

Similaire à An expert system for automatic reading of a text written in standard arabic (20)

SYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMAR
SYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMARSYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMAR
SYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMAR
 
SYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMAR
SYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMARSYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMAR
SYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMAR
 
SYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMAR
SYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMARSYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMAR
SYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMAR
 
SYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMAR
SYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMARSYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMAR
SYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMAR
 
High Quality Arabic Concatenative Speech Synthesis
High Quality Arabic Concatenative Speech SynthesisHigh Quality Arabic Concatenative Speech Synthesis
High Quality Arabic Concatenative Speech Synthesis
 
Survey On Speech Synthesis
Survey On Speech SynthesisSurvey On Speech Synthesis
Survey On Speech Synthesis
 
Emotional telugu speech signals classification based on k nn classifier
Emotional telugu speech signals classification based on k nn classifierEmotional telugu speech signals classification based on k nn classifier
Emotional telugu speech signals classification based on k nn classifier
 
Emotional telugu speech signals classification based on k nn classifier
Emotional telugu speech signals classification based on k nn classifierEmotional telugu speech signals classification based on k nn classifier
Emotional telugu speech signals classification based on k nn classifier
 
Ey4301913917
Ey4301913917Ey4301913917
Ey4301913917
 
Direct Punjabi to English Speech Translation using Discrete Units
Direct Punjabi to English Speech Translation using Discrete UnitsDirect Punjabi to English Speech Translation using Discrete Units
Direct Punjabi to English Speech Translation using Discrete Units
 
PUNJABI SPEECH SYNTHESIS SYSTEM USING HTK
PUNJABI SPEECH SYNTHESIS SYSTEM USING HTKPUNJABI SPEECH SYNTHESIS SYSTEM USING HTK
PUNJABI SPEECH SYNTHESIS SYSTEM USING HTK
 
Development of text to speech system for yoruba language
Development of text to speech system for yoruba languageDevelopment of text to speech system for yoruba language
Development of text to speech system for yoruba language
 
EFFECT OF DYNAMIC TIME WARPING ON ALIGNMENT OF PHRASES AND PHONEMES
EFFECT OF DYNAMIC TIME WARPING ON ALIGNMENT OF PHRASES AND PHONEMESEFFECT OF DYNAMIC TIME WARPING ON ALIGNMENT OF PHRASES AND PHONEMES
EFFECT OF DYNAMIC TIME WARPING ON ALIGNMENT OF PHRASES AND PHONEMES
 
Tutorial - Speech Synthesis System
Tutorial - Speech Synthesis SystemTutorial - Speech Synthesis System
Tutorial - Speech Synthesis System
 
Effect of Dynamic Time Warping on Alignment of Phrases and Phonemes
Effect of Dynamic Time Warping on Alignment of Phrases and PhonemesEffect of Dynamic Time Warping on Alignment of Phrases and Phonemes
Effect of Dynamic Time Warping on Alignment of Phrases and Phonemes
 
E0502 01 2327
E0502 01 2327E0502 01 2327
E0502 01 2327
 
H010625862
H010625862H010625862
H010625862
 
Acceptance Testing Of A Spoken Language Translation System
Acceptance Testing Of A Spoken Language Translation SystemAcceptance Testing Of A Spoken Language Translation System
Acceptance Testing Of A Spoken Language Translation System
 
F017163443
F017163443F017163443
F017163443
 
Implementation of English-Text to Marathi-Speech (ETMS) Synthesizer
Implementation of English-Text to Marathi-Speech (ETMS) SynthesizerImplementation of English-Text to Marathi-Speech (ETMS) Synthesizer
Implementation of English-Text to Marathi-Speech (ETMS) Synthesizer
 

Dernier

How to write a Business Continuity Plan
How to write a Business Continuity PlanHow to write a Business Continuity Plan
How to write a Business Continuity PlanDatabarracks
 
What is DBT - The Ultimate Data Build Tool.pdf
What is DBT - The Ultimate Data Build Tool.pdfWhat is DBT - The Ultimate Data Build Tool.pdf
What is DBT - The Ultimate Data Build Tool.pdfMounikaPolabathina
 
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024BookNet Canada
 
SIP trunking in Janus @ Kamailio World 2024
SIP trunking in Janus @ Kamailio World 2024SIP trunking in Janus @ Kamailio World 2024
SIP trunking in Janus @ Kamailio World 2024Lorenzo Miniero
 
unit 4 immunoblotting technique complete.pptx
unit 4 immunoblotting technique complete.pptxunit 4 immunoblotting technique complete.pptx
unit 4 immunoblotting technique complete.pptxBkGupta21
 
TeamStation AI System Report LATAM IT Salaries 2024
TeamStation AI System Report LATAM IT Salaries 2024TeamStation AI System Report LATAM IT Salaries 2024
TeamStation AI System Report LATAM IT Salaries 2024Lonnie McRorey
 
WordPress Websites for Engineers: Elevate Your Brand
WordPress Websites for Engineers: Elevate Your BrandWordPress Websites for Engineers: Elevate Your Brand
WordPress Websites for Engineers: Elevate Your Brandgvaughan
 
Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 365Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 3652toLead Limited
 
DevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platformsDevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platformsSergiu Bodiu
 
Hyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdf
Hyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdfHyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdf
Hyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdfPrecisely
 
Digital Identity is Under Attack: FIDO Paris Seminar.pptx
Digital Identity is Under Attack: FIDO Paris Seminar.pptxDigital Identity is Under Attack: FIDO Paris Seminar.pptx
Digital Identity is Under Attack: FIDO Paris Seminar.pptxLoriGlavin3
 
Use of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptx
Use of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptxUse of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptx
Use of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptxLoriGlavin3
 
DSPy a system for AI to Write Prompts and Do Fine Tuning
DSPy a system for AI to Write Prompts and Do Fine TuningDSPy a system for AI to Write Prompts and Do Fine Tuning
DSPy a system for AI to Write Prompts and Do Fine TuningLars Bell
 
SALESFORCE EDUCATION CLOUD | FEXLE SERVICES
SALESFORCE EDUCATION CLOUD | FEXLE SERVICESSALESFORCE EDUCATION CLOUD | FEXLE SERVICES
SALESFORCE EDUCATION CLOUD | FEXLE SERVICESmohitsingh558521
 
What's New in Teams Calling, Meetings and Devices March 2024
What's New in Teams Calling, Meetings and Devices March 2024What's New in Teams Calling, Meetings and Devices March 2024
What's New in Teams Calling, Meetings and Devices March 2024Stephanie Beckett
 
Are Multi-Cloud and Serverless Good or Bad?
Are Multi-Cloud and Serverless Good or Bad?Are Multi-Cloud and Serverless Good or Bad?
Are Multi-Cloud and Serverless Good or Bad?Mattias Andersson
 
Passkey Providers and Enabling Portability: FIDO Paris Seminar.pptx
Passkey Providers and Enabling Portability: FIDO Paris Seminar.pptxPasskey Providers and Enabling Portability: FIDO Paris Seminar.pptx
Passkey Providers and Enabling Portability: FIDO Paris Seminar.pptxLoriGlavin3
 
How AI, OpenAI, and ChatGPT impact business and software.
How AI, OpenAI, and ChatGPT impact business and software.How AI, OpenAI, and ChatGPT impact business and software.
How AI, OpenAI, and ChatGPT impact business and software.Curtis Poe
 
Nell’iperspazio con Rocket: il Framework Web di Rust!
Nell’iperspazio con Rocket: il Framework Web di Rust!Nell’iperspazio con Rocket: il Framework Web di Rust!
Nell’iperspazio con Rocket: il Framework Web di Rust!Commit University
 

Dernier (20)

How to write a Business Continuity Plan
How to write a Business Continuity PlanHow to write a Business Continuity Plan
How to write a Business Continuity Plan
 
What is DBT - The Ultimate Data Build Tool.pdf
What is DBT - The Ultimate Data Build Tool.pdfWhat is DBT - The Ultimate Data Build Tool.pdf
What is DBT - The Ultimate Data Build Tool.pdf
 
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
 
SIP trunking in Janus @ Kamailio World 2024
SIP trunking in Janus @ Kamailio World 2024SIP trunking in Janus @ Kamailio World 2024
SIP trunking in Janus @ Kamailio World 2024
 
unit 4 immunoblotting technique complete.pptx
unit 4 immunoblotting technique complete.pptxunit 4 immunoblotting technique complete.pptx
unit 4 immunoblotting technique complete.pptx
 
TeamStation AI System Report LATAM IT Salaries 2024
TeamStation AI System Report LATAM IT Salaries 2024TeamStation AI System Report LATAM IT Salaries 2024
TeamStation AI System Report LATAM IT Salaries 2024
 
WordPress Websites for Engineers: Elevate Your Brand
WordPress Websites for Engineers: Elevate Your BrandWordPress Websites for Engineers: Elevate Your Brand
WordPress Websites for Engineers: Elevate Your Brand
 
Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 365Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 365
 
DevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platformsDevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platforms
 
Hyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdf
Hyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdfHyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdf
Hyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdf
 
Digital Identity is Under Attack: FIDO Paris Seminar.pptx
Digital Identity is Under Attack: FIDO Paris Seminar.pptxDigital Identity is Under Attack: FIDO Paris Seminar.pptx
Digital Identity is Under Attack: FIDO Paris Seminar.pptx
 
Use of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptx
Use of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptxUse of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptx
Use of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptx
 
DMCC Future of Trade Web3 - Special Edition
DMCC Future of Trade Web3 - Special EditionDMCC Future of Trade Web3 - Special Edition
DMCC Future of Trade Web3 - Special Edition
 
DSPy a system for AI to Write Prompts and Do Fine Tuning
DSPy a system for AI to Write Prompts and Do Fine TuningDSPy a system for AI to Write Prompts and Do Fine Tuning
DSPy a system for AI to Write Prompts and Do Fine Tuning
 
SALESFORCE EDUCATION CLOUD | FEXLE SERVICES
SALESFORCE EDUCATION CLOUD | FEXLE SERVICESSALESFORCE EDUCATION CLOUD | FEXLE SERVICES
SALESFORCE EDUCATION CLOUD | FEXLE SERVICES
 
What's New in Teams Calling, Meetings and Devices March 2024
What's New in Teams Calling, Meetings and Devices March 2024What's New in Teams Calling, Meetings and Devices March 2024
What's New in Teams Calling, Meetings and Devices March 2024
 
Are Multi-Cloud and Serverless Good or Bad?
Are Multi-Cloud and Serverless Good or Bad?Are Multi-Cloud and Serverless Good or Bad?
Are Multi-Cloud and Serverless Good or Bad?
 
Passkey Providers and Enabling Portability: FIDO Paris Seminar.pptx
Passkey Providers and Enabling Portability: FIDO Paris Seminar.pptxPasskey Providers and Enabling Portability: FIDO Paris Seminar.pptx
Passkey Providers and Enabling Portability: FIDO Paris Seminar.pptx
 
How AI, OpenAI, and ChatGPT impact business and software.
How AI, OpenAI, and ChatGPT impact business and software.How AI, OpenAI, and ChatGPT impact business and software.
How AI, OpenAI, and ChatGPT impact business and software.
 
Nell’iperspazio con Rocket: il Framework Web di Rust!
Nell’iperspazio con Rocket: il Framework Web di Rust!Nell’iperspazio con Rocket: il Framework Web di Rust!
Nell’iperspazio con Rocket: il Framework Web di Rust!
 

An expert system for automatic reading of a text written in standard arabic

  • 1. International Journal on Natural Language Computing (IJNLC) Vol. 3, No.2, April 2014 DOI : 10.5121/ijnlc.2014.3201 01 AN EXPERT SYSTEM FOR AUTOMATIC READING OF A TEXT WRITTEN IN STANDARD ARABIC Tebbi Hanane1 and Azzoune Hamid2 1 LRIA, Option: Representation of Knowledge and systems of Inference, USTHB, Algiers 2 LRIA, Option: Representation of Knowledge and systems of Inference, USTHB, Algiers ABSTRACT In this work we present our expert system of Automatic reading or speech synthesis based on a text written in Standard Arabic, our work is carried out in two great stages: the creation of the sound data base, and the transformation of the written text into speech (Text To Speech TTS). This transformation is done firstly by a Phonetic Orthographical Transcription (POT) of any written Standard Arabic text with the aim of transforming it into his corresponding phonetics sequence, and secondly by the generation of the voice signal which corresponds to the chain transcribed. We spread out the different of conception of the system, as well as the results obtained compared to others works studied to realize TTS based on Standard Arabic. KEYWORDS Engineering knowledge, modelling and representation of vocal knowledge, expert system, (TTS Text To Speech), Standard Arabic, PRAAT. 1. INTRODUCTION The automatic speech generation is a complex task because of the variability intra and interlocutor of the voice signal. In computer science the difficulty of modeling the speech signal is due to the fact that we don’t know till now how to model very well the enormous mass of knowledge and information useful for the signal synthesis. So we have made a choice to use an expert system to model that knowledge to build a robust system which can really read a text written in a language chosen especially in Standard Arabic. To obtain a better organization of our work, we defined our direct aims. We divided our modeling into three essential stages; the signal analysis, the phonetic orthographical transcription (POT) and finally the synthesis of the textual representation written in Standard Arabic. Therefore the finality considered here is that the user can understand the different sentences transcribed and synthesized which will be pronounced with a clear and high quality way. 2. OVERVIEW OF SPEECH SYNTHESIS SYSTEMS At the present time, we can judge that works carried out in the same context of ours are still not really colossal, and this is because of the complexity of the language itself. And if this some works exists, they are based on the same principle as transcribers of the other languages (French, English, etc.) [1], nevertheless, the efforts undertaken are encouraging and open a large window to follow research task in this field.
  • 2. International Journal on Natural Language Computing (IJNLC) Vol. 3, No.2, April 2014 2 − TTS system MBROLA [2] which use the code SAMPA during the stage of transcription, in this case the user must respect the form of the SAMPA code which is not a universal code; − Work of S. BALOUL [3] who represents a concrete example of transcription of words; based on morphological analysis, and on the studies of pauses to generate the pronunciation of the texts. − SYAMSA (SYstème d’Analyse Morphosyntaxique de l’Arabe), realized by SAROH [4]. According to him, "the phonetisation of the Arabic language is based in particularly on the use of lexicons and on a morphological analyzer for the generation of the different forms of a word. In addition, they are the phenomena of interaction among the words (connection, elision, etc.) and the phenomena of assimilation which suggest the uses of phonological rules "[5]. This tool ensures for each word in entry, the root which correspond to it as well as the morphological and phonetic representations of the root. − The GHAZALI [6] project which was carried out within the IRSIT (Institut Régional des Sciences Informatiques et des Télécommunications) of Tunisia, it is based on a transcription work which fits inside the framework of the realization of a TTS system. The characteristic of this system is shown in the use of a set of rules in the emphasis propagation. − SYNTHAR+ [7] which is a tool studied by Z. ZEMIRLI within the NII (the National Institute of Informatics of Algiers), it ensures the transcription step for a TTS system so that it transmits the phonetic representation to the MULTIVOX synthesizer. It should be known that SYNTHAR+ is based on a morphological analysis before realizing the transcription. 3. DESIGN OF THE SYSTEM The uses of the concept of code and the introduction of high levels in analysis (morphological, syntactic, and pragmatic.) makes the transcription task so difficult and requires deepened studies of the language itself. The difference in our work compared to all that exists is in the transcription using graphemes, i.e. the uses of the Arabic characters as basic units to transcribe directly the text, indeed modeled by an expert system. Thus the role of an expert system is then to infer generally rules like if/then. To build this system our work targets the following steps:
  • 3. International Journal on Natural Language Computing (IJNLC) Vol. 3, No.2, April 2014 3 Figure 1. General schema of our system of speech synthesis 3.1. Acquisition It is the first step of our process of voice synthesis; it plays the role of the interface intermediate between the user and the system. During this time the user have to enter his set of phrases to be pronounced by the system. After that the automatic process can begin. 3.2. Pre-processing and standardization of the text In this phase any form of text is transformed to it literal, perfectly disambiguated form [1]. In particular, this module draft problems of formats, dates and hours, abbreviations, numbers, currencies, addresses, emails…etc. It was shown that majority errors of conversion grapheme/phoneme, for the best operational systems came from the proper names and the exceptions that they pose [8]. Here are some examples of the processing carried out in our case: − The replacement of each composed character by its equivalents Example: ‫ا‬ ‫ل‬ − Consultation of the exceptions lexicon to eliminate the special words. − The application of the transcription rules established for the language. This module must be able to process grammatical complexes tasks to identify the rules of transcription which will be used in a considered context. 3.3. Transcription Two approaches are generally used here: − The use of the lexicon: In this case we must assign to each word in input, the pronunciation which corresponds to it without taking in account its context. The speed, flexibility and simplicity are the main advantages of this approach. It has been shown that the majority errors of conversion grapheme/phoneme, for the best operating systems came from the proper names and the exceptions that they
  • 4. International Journal on Natural Language Computing (IJNLC) Vol. 3, No.2, April 2014 4 pose [9]. Exceptions are words that are not read based on well determined rules of writing; because of this they represent the famous example of this approach. Table 1: a few words of exceptions [5] − The use of rules In this approach each grapheme is converted to a phoneme depending on the context using a set of rewriting rules. The main advantage of this approach come out in the ability to model the linguistic knowledge of human beings by a set of rules that can be incorporated in expert systems. Each of these rules has the following form: [Phoneme] = {LG (LeftContext)} + {C (character)} + {RD (Right Context)} Here is a concrete example of transcription rules [11]: # Is a sign of beginning of sentence, $ Is a sign of end of sentence, § Is one end of word, C Is a consonant, V Is a vowel, SC Is a Solar Consonant and LC is a Lunar Consonant [UU] = {CS} + {‫و‬} +{ ُ} [uu]= §+ {} + { ُ‫و‬ } [UU] = {CL} + {‫و‬} +{ ُ} [uu]= $ {‫و‬ } + { ُ } - -When the " ‫و‬ " is preceded by the vowel/ ُ/ and followed by a consonant, we obtain the long vowel [uu]. - When the " ‫و‬ " is preceded by the vowel / ُ/ in final position, we obtain the long vowel [uu]. The second type represents the diphones which are extracted from the set of artificial words and which do not have a meaning, also called logatomes, or words porters. Each of these is used to extract a single diphone i.e. each logatome contains one and only one diphone and this in order to ensure a degree of independence between this diphone and its context. Here are a few examples of logatomes used : C represents a consonant V a vowel # A silence for the beginning or end of word
  • 5. International Journal on Natural Language Computing (IJNLC) Vol. 3, No.2, April 2014 5 3.4. Analysis (the extraction of the characteristics) The aim of the analysis of the voice signal is to extracts the acoustic vectors which will be used in the stage of synthesis follows. In this step the voice signal is transformed into a sequence of acoustic vectors on the way to decrease redundancy and the amount of data to be processed. And then a spectral analysis by the discrete Fourier transform is performed on a signal frame (typically of size 20 or 30 ms) [9]. In this frame, the voice signal is considered to be sufficiently stable and we extract a vector of parameters considered to be enough for the good operation of the voice signal. In the speech synthesis, the characteristics extraction step, commonly known as the step of analysis, can be achieved in several ways. Indeed, the acoustic vectors are usually extracted using methods such as temporal encoding predictive linear (Linear Predictive Coding LPC) or Cepstrales methods as the MFCC encoding (Mel Frequency Cepstral Coding), as well as the process of segmentation, etc. This process delimits on the acoustic signal a set of segments characterized by labels belonging to the phonetics alphabet of the language under consideration. At present, the segmentation completely automatic of a voice signal remains a fastidious task. Indeed, looking at the complexity acoustico-phonetics phenomena being studied, this activity requires often a manual intervention. Generally, the methods that perform the segmentation of the acoustic wave are divided into two great classes: − The first include all the methods which allow segmenting a voice signal without a priori knowledge of the linguistics content of this signal. These methods split the voice signal into a set of zones homogeneous and stable; − The second class includes all the methods which allow segmenting the voice signal basing on a priori linguistic description (typically on phonemes) of this signal. These methods of segmentation are revealed like methods with linguistically constraints. In our state, we have opted for the second technique using a tool of voice signal analysis which is PRAAT [10] so to slice manually the speech signal in a succession of segments, each one associated with an element acoustic unit (phoneme or diaphone).
  • 6. International Journal on Natural Language Computing (IJNLC) Vol. 3, No.2, April 2014 6 Figure 2. Decomposition in polysound of the word [ ] 3.5. Creation of the sound database Majority of work curried out in the field of the spoken communication required often the recording, and the handling of corpuses of continuous speech, and that to curry out studies on the contextual effects, one the phonetic indices, and variability intra and inter-speaker. Our corpus is modeled by the diagram of class1 as follow: Figure 3. The diagram of class of our sound database 3.6. Voice generation This step commonly known as speech synthesis consist of a transformation of the phonetic sequence (which represents the pronunciation of the written text) resulting from the 1: It is a diagram used in UML(Unified Modeling Language) which is used in object oriented modeling
  • 7. International Journal on Natural Language Computing (IJNLC) Vol. 3, No.2, April 2014 7 transcription step to its substance, i.e. to it acoustics realization. During this step we choose from our sound database the units (phonemes, diphones) most suitable to build the sentence to generate, this means that we will create a automatic function of reading, so that in the end, the user has only to listen to the synthetic sentences. To improve the quality of the generated synthesis, we increase each time the sound unit only. The general silhouette of our expert system can be as follow: Figure 4. General diagram of our expert system In our case we have used the method of synthesis by concatenation and here is our reading function: position= find (grapheme[ ig],API ) ; if((grapheme[ig]==‘‫&&)'أ‬ (grapheme[ig+1]==‘‫))'ل‬ { mp2->FileName="C:son_hananealif.wav"; mp2->Open(); mp2->Wait=true; mp2->Play(); ig=ig+2; position=find (grapheme[ig],API); if(API[position][1]=='L') { mp2->FileName="C:son_hananel.wav"; mp2->Open(); mp2->Wait=true; mp2->Play(); mp2->FileName=API[position][2]; mp2->Open(); mp2->Wait=true; mp2->Play(); ig++; }
  • 8. International Journal on Natural Language Computing (IJNLC) Vol. 3, No.2, April 2014 8 else { if(API[position][1]=='S') { mp2->FileName=API[position][2]; mp2->Open(); mp2->Wait=true; mp2->Play(); ig++; } } } 4. TESTS AND RESULTS To test the performances of our TTS system based on Standard Arabic language, we have chosen a set of sentences which we judged like reference since they contain almost the different possible combinations specific to the language itself. To calculate the success rate (SR) associated with each sentence tested; we got the following formula: %100* drasestesteNumberofph ononcedrasewellprNumberofph SR = The system present in general a SR of 96 % for the set of the sentences tested. Results obtained are summarized by the following table: Table 2. Rate of success for a sample of selected sentences Majority content POT Synthesis by Phonemes Synthesis by Diphones Short vowels 100% 95% / Long vowels 100% 95% / Solar consonants 100% 97% / Lunar consonants 100% 95% / Isolated words 100% 80% 90% Sentences 100% 75% 85% Numbers 90% 95% 100% Exception Words 100% / / 5. CONCLUSION Synthesis systems from text written in Standard Arabic represent a research area very active but difficult to implement. In order to facilitate this task we have proposed a tool, an expert system for voice generation based on Standard Arabic text has been developed, this system uses a set of sub phonetic elements as the synthesis units to allow synthesis of limited-vocabulary speech of good quality. The synthesis units have been defined after a careful study of the phonetic
  • 9. International Journal on Natural Language Computing (IJNLC) Vol. 3, No.2, April 2014 9 properties of modern Standard Arabic, and they consist of central steady-state portions of vowels, central steady-state portions of consonants, vowel-consonant and consonant-vowel transitions, and some allophones. A text-to-speech system which uses this method has also been explored. The input of the system is usual Arabic spelling with diacritics and/or simple numeric expressions. Synthesis is controlled by several text-to-speech rules within the rule database of the expert system, which are formulated and developed as algorithms more suited for computer handling of the synthesis process. The rules are required for converting the input text into phonemes, converting the phonemes into phonetic units, generating the synthesis units from the phonetic units, and concatenating the synthesis units to form spoken messages. The suitability of the method for synthesizing Arabic has been shown by realizing all its functions on a personal computer and by conducting understandability test on synthesized speech. So, we have detailed the different components which represents the basic blocks of our TTS system based on a written text in Standard Arabic and our modeling of it with the use of an expert system, this tool which is a fruits of the artificial intelligence remain less used in the field of automatic speech processing. This fact encouraged us to explore this world trying to give a little push to the research done in this multidisciplinary field. REFERENCES [1] Tebbi Hanane, « Transcription orthographique phonétique en vue de la synthèse de la parole a partir du texte de l’Arabe », Mémoire de magister en Informatique, Blida, Algérie, 2007 [2] http//:tcts.fpms.ac.be/ synthesis/ Mbrola.html. [3] Baloul, « Développement d’un système automatique de synthèse de la parole à partir du texte arabe standard voyellé ». Thèse de doctorat d’université, Le Mans, France, 27 Mai 2003. [4] SAROH Abdelghani, Base de données lexicales dans un système d'analyse morpho-syntaxique de l'arabe : SYAMSA, Toulouse 3, 1989. [5] Guerti Mhania, Tebbi Hanane, La Conversion Graphèmes Phonèmes En Vue D'une Lecture Automatique de Textes en Arabe Standard, LANIA, Chlef, Algérie, 2007 S. [6] Dichy J. & Hassoun M.O. (1998), Some aspects of the DIINAR-MBC research programme, in Ubaydly A., ed.,1998: 2.8.1-5. [7] SYNTHAR+: Arabic speech synthesis under multivox. (SYNTHAR+: Synthèse vocale arabe sous Multivox.) (English) RAIRO, Tech. Sci. Inf. 17, No. 6, 741-761 (1998). [8] Mehdi Braham, Oumaya Abbes, Maroua Trabelsi, Mehdi Dahmen, Rania Nouaari, « Intelligence artificielle : Diagnostic par système expert, Modèle d’analyse d’acteurs : MACTOR » [9] P. Boula de Mareüil, « Synthèse de la parole à partir de courriers et évaluation de la conversion graphème-phonème ». LIMSI-CNRS http://www.limsi.fr/ Individu/ mareuil/ [10] Paul Boersma and David Weenink, "Praat: doing phonetics by computer" Phonetic Sciences, University of Amsterdam Spuistraat 210, 1012VT Amsterdam The Netherlands. [11] T. Saidane, M. Zrigui, & M. Ben Ahmed, « La transcription orthographique phonétique de la langue Arabe ». RECITAL 2004, Fès, 19-22 Avril 2004.