A Comparison Between Allophone, Syllable, and Diphone Based TTS Systems for Kurdish Language Wafa Barkhoda1, Bahram ZahirAzami1, Anvar Bahrampour2, Om-Kolsoom Shahryari1 1

Department of Computer, University of Kurdistan Sanandaj, Iran 2 Department of Computer, Islamic Azad University Sanandaj, Iran {w.barkhoda, zahir, shahryari.kolsoom}@ieee.org , [email protected]

Abstract- nowadays, concatenative method is used in most modern TTS systems to produce artificial speech. The most important challenge in this method is choosing an appropriate unit for creating a database. This unit must warranty smoothness and high quality speech, and also, creating database for it must take reasonable resources and should be inexpensive. Syllable, phoneme, allophone, and, diphone are usually used as the units in such systems. In this paper, we implemented three synthesis systems for Kurdish language, respectively based on syllable, allophone, and diphone. We compare the quality of the three systems, using subjective tests. Keywords- Speech Synthesis; Concatenative Method; Kurdish TTS System; Allophone; Syllable, and Diphone.

1. INTRODUCTION High quality speech synthesis from the electronic form of text has been a focus of research activities during the last two decades, and it has led to an increasing horizon of applications. To mention a few, commercial telephone response systems, natural language computer interfaces, reading machines for blind people and other aids for the handicapped, language learning systems, multimedia applications, talking books and toys are among the many examples [1]. Most of the existing commercial speech synthesis systems can be classified as either formant synthesizers [2,3] or concatenation synthesizers [4,5]. Formant synthesizers, which are usually controlled by rules, have the advantage of having small footprints at the expense of the quality and naturalness of the synthesized speech [6]. On the other hand, concatenative speech synthesis, using large speech databases, has become popular due to its ability to produce high quality natural speech output [7]. The large footprints of these systems do not present a practical problem for applications where the synthesis engine runs on a server with enough computational power and sufficient storage [7]. Concatenative speech synthesis systems have grown in popularity in recent years. As memory costs have dropped, it has become possible to increase the size of the acoustic inventory that can be used in such a system. The first successful concatenative systems were diphone based [8], with only one diphone unit representing each combination of consecutive phones. An important issue for these systems

was how to select, offline, the single best unit of each diphone for inclusion in the acoustic inventory [9,10]. More recently there has been interest in automation of the process of creating databases and in allowing multiple instances of particular phones or groups of phones in the database, with the selection decided at run time. A new, but related problem has emerged: that of dynamically choosing the most adequate unit for any particular synthesized utterance [11]. The development and application of text to speech synthesis technology for various languages are growing rapidly [12,13]. Designing a synthesizer for a language is largely dependant on the structure of that language. In addition, there can be variations (dialects) particular to geographic regions. Designing a synthesizer requires significant investigation into the language structure or linguistics of a given region. In most languages, widespread researches are done on Text-to-Speech systems and also, in some of these languages commercial versions of system are offered. CHATR [14, 15] and AT&T NEXT GEN [16] are two examples offered in English language. Also, in other languages such as French [17,18], Arabic [4,19,20], Norwegian [21], Korean [22], Greek [23], Persian [24-27], etc, much effort has been done in this field. The area of Kurdish Text-to-Speech (TTS) is still in its infancy, and compared to other languages, there has been little research carried on in this language. To the best of our knowledge, nobody has performed any serious academic research on various branches of Kurdish language processing yet (recognition, synthesis, etc.) [28,29]. Kurdish is one of the Iranian languages, which are a sub category of the Indian-European family [30,31]. The Kurdish phonemics consists of 24 consonants, 4 semi vowels and 6 vowels. Also /‫ح‬/, /‫ع‬/, and /‫غ‬/ have entered Kurdish from Arabic. Also, this language has two scripts: the first one is a modified Arabic alphabet and the second one is a modified Latin alphabet [32,33]. For example “trifa” which means “moon light” in Kurdish, is written as /‫ﺗﺮﻳﻔﻪ‬/ in the Arabic script and as “tirîfe” in the Latin. Whereas both scripts are in use, both of them suffer some problems (e.g., in Arabic script the phoneme /i/ is not written; also both /w/ and /u/ are written with the same Arabic written sign /‫و‬/ [32,33], and Latin script does not have the Arabic phoneme /‫ﺋ‬/, and it does not have any standard written sign for foreign phonemes [33]).

In concatenative systems, one of the most important challenges is to select an appropriate unit for concatenation. Each unit has its own advantages and disadvantages, and might be appropriate for a specific system. In this paper we develop three various concatenative TTS systems for Kurdish language based on syllable, allophone, and diphones, and compare these systems in intelligibility, naturalness, and overall quality. The rest of the paper is organized as follows: Section 2 introduces the allophone based TTS system. Section 3 and 4 presents syllable and diphone based systems respectively, and finally, comparison between these systems and quality test results are presented in Section 5. Conclusions are drawn in Section 6. 2. ALLOPHONE BASED TTS SYSTEM In this part, a Text-To-Speech system for Kurdish language, which is constructed based on concatenation method of speech synthesis and use allophones (several pronunciation of a phoneme [34]) as basic unit will be introduced [28,29]. According to the input text, proper allophones from database have been chosen and concatenated to obtain the primary output. Differences between allophones in Kurdish language are normally very clear; therefore, we preferred to explicitly use allophone units for the concatenative method. Some of allophones obey obvious rules; for example if a word end with a voiced phoneme, the phoneme would lose the voicing feature and is called devoiced [32]. However, in most cases there is not a clear and constant rule for all of them. As a result, for extracting allophones we used a neural network. Because their learning power, neural networks can learn from a database and can recognize allophones properly. Fig. 1 shows the architecture of the proposed system. It is composed of three major components: a pre-processing module, a neural network module and an allophone-to-sound module. After converting the input raw text to the standard text, a sliding window of width of four is used as the network input. The network detects second phoneme’s allophone, and the allophone waveform is concatenated to the preceding waveform.

Finally, in standard script, 41 standard symbols were spotted. Notice that this is more than the number of Kurdish phonemes, because we also include three standard symbols for space, comma and dot. Table 1 shows all the standard letters that are used by the proposed system. Table 2 shows the same sentence in various scripts. Also, the standard converter performs standard text normalization tasks such as converting digits into their word equivalents, spelling out some known abbreviations, etc. In the next stage, allophones are extracted from the standard text. This task is done using a neural network. Kurdish phonemes have about approximately 200 allophones, but some of them are very similar, and non-expert people can not detect the differences [32]. As a result, it is not necessary for our TTS system to include all of them (for simplicity, only 66 important and clear instances have been included; see Table 3). Also the allophones are not divided equally between all phonemes (e.g., /p/ is presented by five allophones but /r/ has only one allophone [32]). However, the neural network implementation is very flexible as it is very simple to change the number of allophones or phonemes. Major Kurdish allophones (more than 80% of them) are dependent only on two consecutive phonemes. Others (about 20%) may be dependent on the current, one preceding and two succeeding phonemes [32]. Hence, we employed four sets of neurons in the input layer, each having 41 neurons for detection of the 41 mentioned standard symbols. A sliding window of width four provides input phonemes for the network input layer. Each set of input layer is responsible for one of the phonemes in the window. The aim is to recognize the relevant allophone to the second phoneme of the window. The output layer has 66 neurons (corresponding to the 66 Kurdish allophones used here) for the recognition of the corresponding allophones and the middle layer is responsible for detecting language rules and it has 60 neurons (these values are obtained empirically); (See Fig. 2). The neural network accuracy rate is equal to 98%. In Table 4, neural network output and desired output are compared.

Fig. 1. Architecture of the proposed Kurdish TTS system

The pre-processing module includes a text normalizer and a standard converter. The text normalizer is an application that converts the input text (in Arabic or Latin script) to our standard script; in this conversion we encountered some problems [30-33].

Fig. 2. The neural network structure

After allophone recognition, corresponding waveform of allophones should be concatenated. For each allophone we selected a suitable word and recorded it in a noiseless environment. Separation of allophones in waveforms was

done manually by using of Wavesurfer software. The results of this system and comparison between it and other systems are presented in Section 5. 3. SYLLABLE BASED TTS SYSTEM Syllable is another unit, which is used for developing a text-to-speech system. Various languages have different patterns for syllable. In most of these languages, there are many patterns for syllable and therefore, the number of syllables is large; so usually syllable is not used in allpurpose TTS systems. For example, there are more than 15000 syllables in English [6]. Creating a database for this number of units is a very difficult and time-consuming task. In some languages, the number of syllable patterns is limited, so the number of syllables is small, and creating a database for them is reasonable; therefore this unit can be used in all-purpose TTS systems. For example, Indian language has CV, CCV, VC, and CVC syllable patterns, and the total number of syllables in this language is 10000. In [35], some syllable-like units are used; the number of this unit is 1242. Syllable is used in some Persian TTS systems, too [26]. This language has only CV, CVC, and CVCC patterns for its syllables and so, its syllable number is limited to 4000 [26]. Kurdish has three main groups of syllables that are Asayi, Lekdraw, and Natewaw [33]. Asayi is the most important group and it includes most of the Kurdish syllables. In Lekdraw group, two consonant phonemes occur at the onset of syllable. For example, in /pşû/ two phonemes /p/ and /ş/ make a cluster and the syllable pattern is CCV. Finally, Natewaw group occurs seldom, too. Each group is divided into three groups, Suk, Pir, and Giran [33]. Table 5 shows these groups with corresponding patterns and appropriate examples. According to Table 5, Kurdish has 9 syllable patterns; but two groups Lekdraw and Natewaw are used seldom and in practice, three patterns, CV, CVC, and CVCC are the most used patterns in Kurdish language. According to this fact, we can consider only Asayi group in the implementations, and so the number of database syllables are less than 4000. In our system, we consider these syllables and extend our TTS system using them. 4. DIPHONE BASED TTS SYSTEM Nowadays diphone is the most popular unit in synthesis systems. Diphones include a transition part from the first unit to the next unit, and so, have a more desirable quality rather than other units. Also, in some modern systems, a combination of this unit and other methods such as unit selection are used. Kurdish has 37 phonemes, so as an upper hand estimate, it has 37×36=1332 diphones. However, all of these combinations are not valid. For example, in Kurdish two phonemes /x/ and /g/ or /x/ and /ĥ/ do not succeed each other

immediately. Also, vowels do not form a cluster. So, the number of serviceable diphones in Kurdish is less than 1300. After choosing the appropriate unit, we should choose the suitable instance for each unit. For this reason, we chose a proper word for each diphone and then extract its corresponded signal using COOL EDIT. Quality testing results are discussed in Section 5. 5. QUALITY TESTING RESULTS For evaluating our proposed TTS systems and comparing them, various tests have been carried out. In the first test, a set of seven sentences which were produced with each system was used as the test material. The test sets were played to 17 volunteer listeners. All of the listeners were Kurd and have not any audition problem. The listeners were asked to rate the systems’ naturalness and overall voice quality on a scale of 1 (bad) to 5 (good). The volunteers didn't know anything about the sentences before listening to them. The obtained test results are shown in Table 6. To determine the system's intelligibility, a second test has been carried out. In this test, the listeners were asked to write down the text they understood; then WR and SR were computed using the following equations: WR =

Correct Words Number Correct Syllable Number , SR = Total Words Number Total Syllable Number

Table 7 shows the results for various systems. According to these results, all systems' intelligibilities (especially that of the diphone based system) are acceptable. In the next stage, the Diagnostic Rhyme Test has been used to compare system's quality. The DRT, introduced by Fairbanks in 1958, uses a set of isolated words to test for consonant intelligibility in initial position [36,37]. The test consists of 96 word pairs that differ by a single acoustic feature in the initial consonant. Word pairs are chosen to evaluate the six phonetic characteristics listed in Table 8. The listener hears one word at the time and marks to the answering sheet which one of the two words he thinks is correct. Finally, the results are summarized by averaging the error rates from the answer sheets. For this reason, we chose 96 sets of two one-syllable words; these words are shown in Table 9. Most of these words are meaningful in Kurdish; however, the few meaningless words are shown with bold style. Hence in this test, evaluating the syllable based system quality may show unfairly good results, and we decided not to carry out this test for the syllable based system. The test sets were played to 12 volunteer listeners. Ten listeners were Kurd and two of them were non-Kurd. Table 10 shows the results of this test. The last test which has been carried out is Modified Rhyme Test. The MRT, which is a sort of extension to the DRT, tests for both initial and final consonant apprehension [36,37]. The test consists of 50 sets of six one-syllable words which makes a total set of 300 words. The set of six words is played one at the time and the listener marks which word he

and inexpensive. For example, syllable, phoneme, allophone, and, diphone are usually considered as appropriate for allpurpose systems. In this paper, we implemented three synthesis systems for Kurdish language based on syllable, allophone, and diphone and compared their quality using various tests. The diphonebased TTS system showed to be the most natural one while all systems' Intelligibilities are acceptable. Unit selection method [40] can produce high quality and natural output speech. Developing a TTS system using unit selection and combining it with other methods is our goal in the future works.

thinks he hears on a multiple choice answer sheet. The first half of the words is used for the evaluation of the initial consonants and the second one for the final ones. Table 11 summarizes the test format [38]. Results are summarized as in DRT, but both final and initial error rates are given individually [39]. The same group of 12 listeners as in our DRT test has been used in this test. Final results are shown in Table 12. 6. CONCLUSION AND FUTURE WORKS Nowadays, most modern TTS systems use concatenative method to produce artificial speech. The most important challenge in this method is choosing appropriate unit for the database. This unit must warranty smoothness and high quality speech, and creating the database must be reasonable

Arabic Latin Standard

‫غ‬ X ‫ﯼ‬ y y

Arabic Latin Standard

‫ع‬ G

‫ش‬ Ş S

‫ﯼ‬ î I

‫س‬ s s

‫ﻩ‬ e e

i i

Table 1: ‫ژ‬ j j ‫ێ‬ ê Y

Arabic Format Latin Format Standard Format

List of the proposed system's standard letters ‫ز‬ ‫ڕ‬ ‫ر‬ ‫د‬ ‫خ‬ ‫ح‬ ‫چ‬ ‫ج‬ z rr r d x ç c z R R d x H C c

‫ه‬ h h

‫وو‬ û U

‫ۆ‬ o o

‫و‬ U U

‫و‬ w w

‫ن‬ n n

‫م‬ m m

‫ڵ‬ ll L

‫ت‬ t t ‫ل‬ l l

‫پ‬ p p

‫گ‬ g g

‫ب‬ b b

‫ﮎ‬ k k

‫ا‬ a a

‫ق‬ q q

‫ﺋ‬ A

‫ڤ‬ v v

‫ف‬ f f

Table 2: The same sentence in various scripts ‫دﻟﻮپ دﻟﻮپ ﺑﺎران ﮔﻮل ﺋﻪ ﻧﻮوﺳﻴﺘﻪ وﻩ و ﻧﻤﻪ ﻧﻤﻪ ﻳﺶ ﭼﺎواﻧﻢ ﺗﻮ‬ Dillop dillop baran gull enûsêtewe û nime nimeyş çawanim to diLop diLop baran guL AenUsYtewe U nime nimeyS Cawanim to

Phoneme Allophones

Table 3: List of the phonemes and their corresponding allophones as used in the proposed system P b t d K g Q F s S z J PpO*& bEB t@T d!D k?K G%g Qq FVf s $S zZ> Jj

G ^

A A

Phoneme Allophones

C Cc

R R

Y Y

Phoneme Allophones

h h

H H E E

a a

m mWM

X X

N #

U U

X X u u

n nN o o

v v

y y

w w

l l

I I

i i

r r C ~

L L Ü _

Table 4: A comparison between neural network output and desired output NN Output DiLo&_DiLoP_baraN_GuL_AenUsY@_U,_nime_nimeyS_~awaniM_to Desired Output DiLo&_DiLo&_baraN_GuL_AenUsY@_U,_nime_nimeyS_~awaniM_to

Table 5: Kurdish syllable patterns Suk Pir Asayi Lekdraw Natewaw

Giran

Syllable Pattern Example Syllable Pattern

CV De, To CCV

CVC Waz, Lîx CCVC

CVCC Kurt, Berd CCVCC

Example

Bro, çya

Bjar, Bzût

Xuast, Bnêsht

Syllable Pattern Example

V -î

VC -an

VCC -and

Table 6: First test results Naturalness Overall Quality Allophone Based System Syllable Based System Diphone Based System

2.29 2.65 3.37

2.45 3.02 3.51

. .

, ,

Table 7: Second test results WR Allophone Based System Syllable Based System Diphone Based System

79.4 82.6 93.8

SR 83.2 87 97.2

Table 8: The DRT characteristics Characteristics Description Example Voicing Voiced - Unvoiced veal - feel Nasality Nasal - Oral reed - deed Sustension Sustained - Interrupted Sheat - cheat Sibilation Sibilated - Unsibilated Sing - thing Graveness Grave – Acute Weed - reed Compactness Compact - Diffuse Key - tea

Table 9: Kurdish minimal pair words used in DRT test

Voicing ban tan gall kall bûll pûll dîn tîn dêr têr zall sall zerd serd gorr korr fîz vîz zuêr suêr birr pirr ders ters gom çom zirr sirr girr çirr vam fam

Nasality mêz têz merd terd nan dan nêz dêz mall tall mîn tîn mil til nas das nem dem maf taf noş doş nêr dêr nûr dûr nerd derd man tan maş taş

Voicing Allophone Based System Syllable Based System Diphone Based System

96.87 --97.91

Sustention şirr çirr ver ber firr pirr şem çem şill çill van ban fil pil fall pall şen çen var bar şall çall şorr çorr şing çing fîs pîs faş paş vor bor

Sibilation çoll koll soz toz sam tam jîr gîr jar yar sall tall zîn tîn sem tem zam tam zil til zem tem çall kall saf taf çem kem çirr kirr sêx têx

Graveness bar dar perr terr berd derd fall tall pek tek pirs tirs fam tam birr dirr boll doll fêr têr pêm têm ball dall ban dan ferz terz fil til pall tall

Table 10: The results of DRT test Nasality Sustension Sibilation 97.39 --98.95

94.27 --95.31

95.83 --96.35

Compactness turd kurd han fan fall hall yar var duê kuê tam kam torr korr dall gall ferd herd tîj kîj derd gerd yall vall tuê kuê dirr girr tall kall tem kem

Graveness

Compactness

Average

95.31 --97.39

97.91 --98.43

96.26 --97.39

Table 11: Examples of the response sets in MRT A B C D E

F

1 2 …

bar korr

ban koll

baş koş

bax kox

barr kot

bas kok

26 27 28 …

ban dall hoz

tan hall toz

man gall soz

yan mall moz

wan fall qoz

şan yall poz

Table 12: The final MRT results Initial position Final Position Allophone Based System Syllable Based System Diphone Based System

77.7 --88.4

81.4 --86.4

Average 79.55 --87.4

REFERENCES [1] H. Al-Muhtaseb1, M. Elshafei1, and M. Al-Ghamdi, "Techniques for High Quality Arabic Speech Synthesis," Information sciences, Elsevier, 2002. [2] T. Styger and E. Keller, Fundamentals of Speech Synthesis and Speech Recognition: Basic Concepts, State of the Art, and Future Challenges Formant synthesis, In Keller E. (ed.), 109-128, Chichester: John Wiley, 1994. [3] D. H. Klatt, "Software for a Cascade/Parallel Formant Synthesizer,” Journal of the Acoustical Society of America, Vol 67, 971-995, 1980. [4] W. Hamza, Arabic Speech Synthesis Using Large Speech Database, PhD. thesis, Cairo University, Electronics and Communications Engineering Department, 2000. [5] R. E. Donovan , Trainable Speech Synthesis, PhD. thesis, Cambridge University, Engineering Department, 1996. [6] S. Lemmetty, Review of Speech Synthesis Technology, M.Sc Thesis, Helsinki University of Technology, Department of Electrical and Communications Engineering, 1999. [7] A. Youssef , et al, "An Arabic TTS System Based on the IBM Trainable Speech Synthesizer,” Le traitement automatique de l’arabe, JEP–TALN 2004, Fès, 2004. [8] J. P. Olive, "Rule synthesis of speech from diadic units," ICASSP, pages 568-570, 1977. [9] A. Syrdal, "Development of a female voice for a concatenative textto-speech synthesis system," Current Topics in Acoust. Res., 1:169181, 1994. [10] J. Olive, J. van Santen, B. Moebius, and C. Shih, Multilingual Textto-Speech Synthesis: The Bell Labs Approach, pages 191-228. Kluwer Academic Publishers, Norwell, Massachusetts, 1998. [11] M. Beutnagel, A. Conkie, and A. K. Syrdal, "Diphone Synthesis using Unit Selection," The Third ESCA/COCOSDA Workshop (ETRW) on Speech Synthesis, ISCA – 1998. [12] R. Sproat, J. Hu, and H. Chen, "Emu: An e-mail pre-processor for text-to-speech," Proc. IEEE Workshop on Multimedia Signal Proc., pp. 239–244, Dec. 1998. [13] C. H. Wu and J. H. Chen, "Speech Activated Telephony Email Reader (SATER) Based on Speaker Verification and Text-toSpeech Conversion," IEEE Trans. Consumer Electronics, vol. 43, no. 3, pp. 707- 716, Aug. 1997. [14] A. Black, CHATR, Version 0.8, a generic speech synthesis, System documentation, ATR-Interpreting Telecommunications Laboratories, Kyoto, Japan, March 1996. [15] A. Hunt and A. Black, "Unit selection in a concatenative speech synthesis system using a large speech database," ICASSP, 1:373376, 1996. [16] M. Beutnagel, A. Conkie, J. Schroeter, Y. Stylianou, and A. Syrdal, "The AT&T NEXT-GEN TTS System," Joint Meeting of ASA, EAA, and DAGA, 1999. [17] T. Dutoit, High Quality Text-To-Speech Synthesis of the French Language, Ph.D. dissertation, submitted at the Faculté Polytechnique de Mons, 1993. [18] T. Dutoit, et al, "The MBROLA project: towards a set of high quality speech synthesizers free of use of non commercial purposes," ICSLP 96, Proceedings, Fourth International Conference, IEEE, 1996. [19] F. Chouireb, M. Guerti, M. Naïl, and Y. Dimeh, "Development of a Prosodic Database for Standard Arabic," Arabian Journal for Science and Engineering, 2007. [20] A. Ramsay and H. Mansour, "Towards including prosody in a text-tospeech system for modern standard Arabic," Computer Speech & Language, Elsevier, 2008. [21] I. Amdal and T. Svendsen, "A Speech Synthesis Corpus for Norwegian", lrec'06, 2006. [22] K. Yoon, "A prosodic phrasing model for a Korean text-to-speech synthesis system," Computer Speech & Language, Elsevier, 2006. [23] P. Zervas, I. Potamitis, N. Fakotakis, G. Kokkinakis, "A Greek TTS based on Non uniform unit concatenation and the utilization of

Festival architecture," First Balkan Conference on Informatics, Thessalonica, Greece, pp. 662-668, 2003. [24] A. Farrohki, S. Ghaemmaghami, and M. Sheikhan, "Estimation of Prosodic Information for Persian Text-to-Speech System Using a Recurrent Neural Network," ISCA, Speech Prosody 2004, International Conference, 2004. [25] M. Namnabat and M. M. Homayunpoor, "Letter-to-Sound in Persian Language Using Multy Layer Perceptron Neural Network," Iranian Electrical and Computer Engineering Journal, 2006 (in Persian). [26] H. R. Abutalebi and M. Bijankhan, "Implementation of a Text-toSpeech System for Farsi Language," Sixth International Conference on Spoken Language Processing (ISCA), 2000. [27] F. Hendessi, A. Ghayoori, and T. A. Gulliver, "A Speech Synthesizer for Persian Text Using a Neural Network with a Smooth Ergodic HMM," ACM Transactions on Asian Language Information Processing (TALIP), 2005. [28] F. Daneshfar, W. Barkhoda, and B. ZahirAzami, "Implementation of a Text-to-Speech System for Kurdish Language," ICDT'09, Colmar, France, July 2009. [29] W. Barkhoda, F. Daneshfar, and B. ZahirAzami, "Design and Implementation of a Kurdish TTS System Based on Allophones Using Neural Network," ISCEE'08, Zanjan, Iran, 2008 (In Persian). [30] W. M. Thackston, Sorani Kurdish: A Reference Grammar with Selected Reading, Harvard: Iranian Studies at Harvard University, 2006. [31] A. Rokhzadi, Kurdish Phonetics and Grammar, Tarfarnd press, Tehran, Iran, 2000 (In Persian). [32] M. Kaveh, Kurdish Linguistic and Grammar (Saqizi accent), Ehsan Press, first edition, Tehran, ISBN 964-356-355-3, 2005 (In Persian). [33] S. Baban, Phonology and Syllabication in Kurdish Language, Kurdish Academy Press, first edition, Arbil, 2005 (In Kurdish). [34] R. J. Deller, et al., Discrete time processing of speech signals, John Wiley and Sons, 2000. [35] M. N. Rao, S. Thomas, T. Nagarajan, and H. A. Murthy, "Text-toSpeech Synthesis using syllable-like units," National Conference on Communication, India, 2005. [36] M. Goldstein, "Classification of Methods Used for Assessment of Text-to-Speech Systems According to the Demands Placed on the Listener," Speech Communication, vol. 16: 225-244, 1995. [37] J. Logan, B. Greene, and D. Pisoni, "Segmental Intelligibility of Synthetic Speech Produced by Rule," Journal of the Acoustical Society of America, JASA vol. 86 (2): 566-581, 1989. [38] Y. Shiga, Y. Hara, and T. Nitta, "A Novel Segment-Concatenation Algorithm for a Cepstrum-Based Synthesizer," Proceedings of ICSLP 94, (4): 1783-1786, 1994. [39] D. Pisoni and S. Hunnicutt, "Perceptual Evaluation of MITalk: The MIT Unrestricted Text-to-Speech System," Proceedings of ICASSP 80, vol. 5: 572-575, 1980. [40] H. Sak, A Corpuse-Based Concatenative Speech Synthesis System for Turkish, M.Sc. Thesis, Bogazici University, 2004.

A Comparison Between Allophone, Syllable, and ...

done manually by using of Wavesurfer software. The results of this system and .... dêr têr mall tall şill çill jar yar pek tek duê kuê zall sall mîn tîn van ban sall tall ... [7] A. Youssef , et al, "An Arabic TTS System Based on the IBM. Trainable Speech ...

213KB Sizes 2 Downloads 229 Views

Recommend Documents

Absolute and relative signals: a comparison between melanin- and ...
Summary. Some signalling traits may be highly influenced by environmental factors, thereby decreasing between-year trait repeatability what could compromise the honesty of signal. However, although the repeatability of environmentally-influenced trai

Comparison between Biexponential and Robust Biexponential ...
Comparison between Biexponential and Robust Biexponential Nonlinear Models, Using Simulation.pdf. Comparison between Biexponential and Robust ...

Modesty in selfpresentation: A comparison between the USA and Japan
Modesty in self-presentation: A comparison between the USA and Japanajsp_ 60..68. Toshio Yamagishi,1 Hirofumi Hashimoto,1 Karen S. Cook,2 Toko Kiyonari,3. Mizuho Shinada,1 Nobuhiro Mifune,1 Keigo Inukai,1 Haruto Takagishi,1 Yutaka Horita1 and Yang Li

A Comparison Between Broad Histogram and ... - Springer Link
KEY WORDS: Microcanonical averages; numerical simulation. We could find three ... choose the energy E0 at the center of the spectrum, and equate our numerically .... authors re-formulated their criticism concerning what they call ''systemati-.

A comparison between 3G and 802.11 wireless ...
A comparison between 3G and 802.11 wireless technologies for Inter- ... Vehicular Ad Hoc Network, a VANET, between the ..... coverage. In order to guarantee no interference of any kind, the decision was taken to test each technology ...

A Comparison Between Broad Histogram and ... - Springer Link
called Entropic Sampling, which, from now on, we call ESM. We present ... to note that these movements are virtual, since they are not actually per- ..... B. A. Berg, in Proceedings of the International Conference on Multiscale Phenomena and.

economics of the internet: a comparison between
using the number of Internet hosts for Western Europe as dependent ... (HOST), telephone lines per 1000 people (TELLINE), rural population as a ..... countries, SMS through mobile phones has become affordable and a popular mode of ... Hargittai, E.,

A comparison of ground geoelectric activity between three regions of ...
A comparison of ground geoelectric activity between three regions of different level of ..... To go further inside in the comparison of our data sets, we constructed ...

Multilineage Differentiation Capability Comparison between ... - hikari
MSCs were observed under the electron microscopes. Scanning .... the process of human development” [Verfaillie et al., 2003] urges us to reexamine the.

A comparison between concept maps, mind maps ...
without drawbacks 15–17 and they may not fit all types of target groups (such as non-academics), learning tasks (i.e. developing .... There is a myriad of node-link mapping methods from such diverse areas as psychology, computer .... and facilitati

A Comparison between the stochastic intensity SSRD ...
Sep 12, 2004 - We introduce an analytical approximation for the SSRD im- ..... not necessary to store all the paths. ..... 6.1 Market data and Simulation Setup.

A comparison of ground geoelectric activity between three regions of ...
ing exponents for short and large lags arisen from crossover points in the geoelectric ... we introduce the method of data processing; in Sect. 4 the re- sults of the ...

20140228 Comparison between MARATHON.pdf
garbage to the course. Strict prohibition!Violators will be immediately disqualified & will be. asked to leave the course. Mostly no Environmental Protection. Rules & Guidelines. If you find any trash on the. course. Be a HERO, by picking it up and t

Multilineage Differentiation Capability Comparison between ... - hikari
After blocked with PBS containing 2% BSA, cells were permeabilized with 0.1% Triton-X 100 for 10 min. Slides were incubated sequentially overnight at 4°C with ...

Comparison between discrete dipole implementations ...
The geometry of the scatterer is read from a file and all the parameters of .... unlimited number of dipoles, since ADDA is not limited by the memory of a single ... symmetry of the interaction matrix is used to decrease storage requirement of its ..

Comparison between discrete dipole implementations ...
in astronomy and in some technological applications has greatly increased in the last years. ..... [16] or the more advanced package 'fastest Fourier transform in the west' (FFTW) [26]. ...... science and is owned by the Ministry of Education.

Comparison of insect biodiversity between organic and conventional ...
Aug 26, 2014 - 3 Centre for Wildlife Studies, CWS, Bengaluru, Karnataka, India. 4 UAS - University of Agricultural Sciences, GKVK Campus, ... data; AG supervised the research plan and implementation of the field work and wrote up the paper. .... Figu

Comparison between Course Options and Youth Options August 2015 ...
There was a problem previewing this document. Retrying. ... Comparison between Course Options and Youth Options August 2015 v1.0.pdf. Comparison ...

Comparison between Position and Posture Recovery in ...
Service System Technology Development for Ubiquitous City). Fig. 1. Three Types of Recovery Methods [2] the robot on the path. It does not consider orientation ...

Syllable Integrity - Dan Everett
Next, I review the basic properties of Banawá stress, as originally analyzed in ... However, if it does have V-syllables, then we must also add a stipulation to the effect .... 10 For the sake of illustration, I have shown the forms of the name and.