ACDSee列印工作

Size: px
Start display at page:

Download "ACDSee列印工作"

Transcription

1

2

3

4

5 Computational Linguistics and Chinese Language Processing Vol. 15, No. 2, June 2010, pp The Association for Computational Linguistics and Chinese Language Processing A Punjabi to Hindi Machine Transliteration System Gurpreet Singh Josan, and Gurpreet Singh Lehal Abstract Transliteration is the general choice for handling named entities and out of vocabulary words in any MT application, particularly in machine translation. Transliteration (or forward transliteration) is the process of mapping source language phonemes or graphemes into target language approximations; the reverse process is called back transliteration. This paper presents a novel approach to improve Punjabi to Hindi transliteration by combining a basic character to character mapping approach with rule based and Soundex based enhancements. Experimental results show that our approach effectively improves the word accuracy rate and average Levenshtein distance of the various categories by a large margin. Keywords: Transliteration, Punjabi, Hindi, Soundex Approach, Rule based Approach, Word Accuracy Rate. 1. Introduction Every machine translation system has to deal with out-of-vocabulary words, like technical terms and proper names of person, places, objects, etc. Machine transliteration is an obvious choice for such words. When words cannot be found in translation resources, such as a bilingual dictionary, transliteration - the process of converting characters in one alphabet into another alphabet - is used. Transliteration is a process wherein an input string in some alphabet is converted to a string in another alphabet, usually based on the phonetics of the original word. If the target language contains all the phonemes used in the source language, the transliteration is straightforward, e.g. the Hindi transliteration of Punjabi word ਕਮਰ [kamarā] (room) 1 is कमर [kamarā], which is essentially pronounced in the same way. Nevertheless, if some of the sounds are missing or are extra in the target language, they are generally mapped to the most phonetically similar letter, e.g., in Hindi we have syllables (consonant clusters) that are written as a single atomic grapheme and have a double sound Department of Computer Science, Punjabi University Patiala, Punjab, India josangurpreet@rediffmail.com; gslehal@gmail.com The author for correspondence is Gurpreet Singh Josan. 1 The text in [ ] denotes phonetic transcription whereas in ( ) denotes the English translation of a word.

6 78 Gurpreet Singh Josan and Gurpreet Singh Lehal associated with them, like क ष, which is a combination of the sound of क and ष. No such consonant cluster is present in Punjabi. So, a similar sounding letter generally is used to denote such sounds. Again, there is no rule to find out when a sequence of letters in Punjabi is going to map in a consonant cluster in Hindi, e.g., consider the two names written in Punjabi, viz. ਸ਼ ਤਜ [ṣitij] क ष तज [kṣitija] and ਸ਼ਕ ਕ ਈ [śikākāī] शक क ई [śikākāī]. In the first name, the consonant cluster ਸ਼ is mapped to क ष, whereas, in the second name, the same is mapped to श. The important consideration in transliteration is to preserve the phonetics of transliterated word (Virga & Khudanpur, 2003). The transliteration process depends on the target language and on the education and experience of the actual transliterator. Thus, a single foreign word can have many different transliterations. For example, ਮ ਹਫ ਜ਼ [mahiphūz] (safe) can be transliterated as महफ ज़ [mahaphuz], महफ़ ज़ [mahafuz], म हफ ज़ [mahiphuz], म हफ़ ज़ [mahifuz], etc. This variation, due to localization, poses problems in various NLP tasks, like cross language information retrieval. According to Viswanadha R. (2002), for a finer transliteration among two scripts, the scripts must have the features of completeness, predictability, pronounceable, unambiguousness, and partial reversibility. Transliteration schemes have to face the problem of letters present in one language and not in the other. Unless a superset of letters from all of the Indian Languages is formed, uniform transliteration is ruled out. Viswanadha, (2002) had the view that, when characters do not have any appropriate transliteration, they should be consumed and not replaced with any other character. This results in partial loss of reversibility and readability. For several decades now, Roman transliteration has been used in English Indian language texts. Extensive research has been carried out on methodologies for transliterating Indian scripts to and from their Romanized counterpart. Transliteration among Indian scripts is a rather neglected area. With the development of ISCII (Indian Script Code for Information Interchange), the problems in transliteration among Indian scripts have been solved to some extent. ISCII has been designed via the phonetic property of Indian scripts and caters to the superset of all Indian scripts. The ISCII document is IS13194:1991, available from the office of the Bureau of Indian Standards. The ISCII approach is based on the fact that many Indian language groups share a common set of phonemes, and via the same set of codes for different Indian languages, it would be possible to change fonts from one language to the other, thereby "transliterating" the script in the process. By attaching an appropriate script rendering mechanism to ISCII, transliteration from one Indian script to another is achieved in a natural way. Later in the decade of the 1990s, Unicode standards were developed for representing the character set of all languages. In Unicode, alphabets are coded by script rather than language, which saves reduplication of the same character if it occurs in multiple languages. The Gurmukhi and Devanagari Unicode tables were based on templates from the 1988 versions of

7 A Punjabi to Hindi Machine Transliteration System 79 the ISCII standard. In this paper, we will discuss the Punjabi to Hindi Machine transliteration system. Although Punjabi and Hindi are related languages, and, except for a few cases, all letters of Gurmukhi script (a script for the Punjabi Language) are present in Devanagri script (a script for the Hindi Language), the task of transliteration from Punjabi to Hindi is not trivial. We will use letter to letter mapping as the baseline and try to find the improvements by rule based and Soundex based approaches. This work forms part of larger Punjabi to Hindi machine translation system, and the results of transliteration will be applied to this system for its improvement. The remainder of the paper is organized as follows. Section 2 reviews previous work. In Section 3, the Gurmukhi and Devanagri scripts are compared, and the problems of transliteration from Punjabi to Hindi are discussed. Section 4 describes the formulation of the rule based model and the Soundex based model. Section 5 outlines the experiment and the evaluation methodology used. Section 6 will present the results. Finally, Section 7 contains our overall impressions and conclusions and points to future work. 2. Previous Work The topic of machine transliteration has been studied extensively for several different language pairs, and many techniques have been proposed. Grapheme based models and Phoneme Based models are the two approaches found in literature. Grapheme refers to the basic unit of a written language: for example English has 26 graphemes or letters. Phonemes are the simplest significant unit of sound e.g. the /M/, /AE/, and /TH/ in math. Grapheme based models are classified into the statistical transliteration based model, decision tree based model, transliterated network based model etc. Some examples of this would be the noisy-channel model (NCM) (Lee & Chang, 2003; Virga et al., 2003), HMM (Jung, Hong & Paek, 2000), decision tree (Kang & Choi, 2000), transformation-based learning (Meng, Lo, Chen & Tan, 2001), statistical machine transliteration model (Lee et al. 2003), finite state transducers (Knight & Grahel, 1998), and rule-based approach (Oh & Choi, 2002; Wan & Verspoor, 1998). The phoneme-based approach has received remarkable attention in various works (Lee et al. 2003; Meng et al. 2001; Oh et al., 2002; Knight et al. 1998; Virga et al. 2003; Jung et al. 2000; Al-Onaizan & Knight, 2002). For Indian languages, as mentioned earlier, Roman transliteration has been used to represent texts of Indian languages in English. Since it is difficult to represent the letters of Hindi using just the twenty six letters of the Roman alphabet, scholars have used varying schemes to accommodate sounds that could not be correctly indicated using appropriate Roman letters. The schemes are somewhat arbitrary in the choice of Roman letters. Sometimes, phonetic symbols are used in place of the normal Roman letters. Most phonetic symbols are

8 80 Gurpreet Singh Josan and Gurpreet Singh Lehal basically the letters of the Roman and Greek alphabets with special marks known as diacritic marks. Roman transliteration that makes use of diacritic marks works better for Indian languages, and some standardization has been effected in the last few decades based on the recommendation from the National Library in Calcutta ( Transliteration Principles, 2008). Roman letter assignments in this scheme are phonetically equivalent to the letters of Sanskrit or other Indian languages. The primary difficulty in data entry of the phonetic symbols is that there is no provision to input the symbols directly using the standard ASCII keyboard. Transliteration methods using only the displayable ASCII symbols do not run into this problem since the ASCII letters can be typed in directly. A special computer program, however, would be required to interpret the input string to produce the Indian languages display or printout. This is precisely what the currently popular transliteration schemes attempt. Schemes such as ITRANS (Chopde, 2001), RIT (Kanneganti & Kishore, 2008), ADHAWIN (Srinivasan, 1995), and MYLAI (KalyanaSundram, 2008), use only the standard displayable ASCII letters and symbols to transliterate the text. These schemes allow multiple representations for certain syllables and long vowels but the processing program handles this well. Although the input can be done using standard keyboards, these schemes are not suitable for searching the contents on the Internet. There is a need to evolve a case-insensitive transliteration scheme to facilitate searching on web. For the Punjabi language, a Gurmukhi to Roman transliteration system using a transliteration scheme based on ISO: transliteration and ALA-LC has been developed at Punjabi University Patiala (Sharma, 2008). All of these transliteration schemes are either from Roman to Indian languages or vice-versa. Transliteration among Indian languages is rather ignored. As the English language is one of the official languages of India, Roman transliteration is developed primarily to disseminate information in the English language. Little need is felt for transliteration among Indian languages. Malik (2006) has developed a machine transliteration system from Hindi to Urdu. The system converts Hindi and Urdu into a common language that may be ASCII encodings of Hindi and Urdu characters or the International Phonetic Alphabet (IPA) equivalents of Hindi and Urdu Phonemes. Then, Hindi and Urdu will be generated from the common language. This makes the system reversible. A corpus based transliteration system for Shahmukhi to Gurmukhi has also been developed (Saini & Lehal, 2008). Their system uses statistical data from both Shahmukhi and Gurmukhi corpora like character, word, and n-gram frequencies. In this system, script mappings are performed for simple consonants, aspirated consonants (AC), vowels, and other diacritical marks or symbols. Next, the transliteration system is virtually divided into two phases. The first phase performs rule-based transliteration, and the second phase performs the transliteration using bi-gram language model.

9 A Punjabi to Hindi Machine Transliteration System 81 A machine transliteration system has also been developed for transliterating from Hindi to Punjabi (Goyal & Lehal, 2009). This system has implemented approximately fifty complex rules for making the transliteration between Hindi-Punjabi language pair. The system claims 98% accuracy. The system is not reversible due to the dependency on language-specific rules. As more and more data in Indian languages is available in electronic format, the need of the hour is to develop sophisticated transliteration modules for Indian languages that can be used in various NLP systems, like machine translation and cross language information retrieval. This paper discusses the first attempt to develop a transliteration system for transliterating Punjabi text to Hindi. 3. Gurmukhi and Devanagari Scripts Punjabi can be written in Gurmukhi script or in Shahmukhi script. Shahmukhi is used to write and record the Punjabi language in West Punjab in Pakistan. In East Punjab in India, however, the Gurmukhi script is used to record the Punjabi language. This paper also considers this script for writing Punjabi. The script used for writing Hindi is called Devanagari. Both Gurmukhi and Devanagari descended from the Brahmi script of Ashoka. Both scripts are left-to-right and words are spelled phonetically. As a background for readers who are not familiar with script terminology, we will describe the basic elements of a script here. Building blocks of a script are alphabets, which are the sets of letters, each of which represents one or more phonemes in the language they are used to write. In some cases, combinations of letters are used to represent single phonemes, as in the English sh, ch, and th. The alphabets contain letters for both consonants (sound articulated with complete or partial closure of the vocal tract) and vowels (sound pronounced with an open vocal tract). A single vowel or a consonant plus a vowel make a syllable. A writing system can be syllabary or alphasyllabary. A syllabary system is a phonetic writing system consisting of symbols representing syllables. On the other hand, an alphasyllabary system consists of symbols for consonants and vowels. The alphasyllabary system is also known as abiguda. Both Gurmukhi and Devanagri are alphasyllabary in nature. Alphabets of both scripts represent syllables. All consonants contain an inherent vowel /a/ or schwa ending, both of which can be altered or muted by means of diacritics or matra. Vowels can also be written with separate letters when they occur at the beginning of a word or on their own. When two or more consonants occur together, special conjunct symbols are often used to add the essential parts of the first letter or letters in the sequence to the final letter ( Writing Systems, 2010). Gurmukhi, meaning "from the mouth of the Guru" is the most commonly used script in India for writing in Punjabi. Gurmukhi was introduced by the second Guru of the Sikhs, Guru Angad Dev Ji, in the sixteenth century as a polished version of the Landa writing system, which was used in old Punjabi at that time. Part of the shift from Landa to Gurmukhi entailed

10 82 Gurpreet Singh Josan and Gurpreet Singh Lehal the inclusion of some Devanagari letters. Gurmukhi was then used to record the scriptures of the Sikhs, and it continues till today as the prominent writing system for Punjabis. Bhatia (1993) identifies three stages of development the Punjabi language has undergone over the years, which are illustrated in Figure 1 below. Stage 1 : Old Punjabi (10th to 16 th century) Stage 2 : Medieval Punjabi (16th to 19th century) Stage 3 : Modern Punjabi (19th century to present) Figure 1. The development stages of the Punjabi language. In Gurmukhi, there are three letters that are used to provide bases for free-standing vowels: ੳ [uṛa] ਅ [ æṛa] ੲ [iṛi] There are nine dependent vowel sounds (also called diacritics) available in Punjabi as,,,,,,,, and. The allowed combination of vowel symbols and sounds are ਅ, ਆ, ਇ, ਈ, ਉ, ਊ, ਏ, ਐ, ਓ, and ਔ. These are also called independent vowels. All of these combinations have a unique code in Unicode and are termed as independent vowels. When a vowel sound comes at the start of a word or is independent of some consonant in the middle or end of a word, independent vowels are used. There are 32 consonants in the older Gurmukhi script. Later, some more letters were introduced to represent some sounds adapted from Farsi and Persian. Presently, Gurmukhi has 38 consonants, 10 vowels letters (independent vowels), 9 vowel symbols (dependent vowels), 2 symbols for nasal sounds, and 1 symbol that duplicates the sound of consonants (Bhatia, 1993; Malik, 2006). The form of each syllable is often dependent on whether it occurs in combination with other letters, e.g., a vowel following a consonant changes its form, behaving as a diacritic modifier to the consonant: ਕ + ਈ = ਕ All consonants have an inherent vowel, i.e. /a/. Diacritics, which can appear above, below, before or after the consonant they belong to, are used to change the inherent vowel. When certain consonants occur together, a special conjunct symbol halant character before the consonant is used to combine the essential parts of each letter. For example ਕ + + ਰ = ਕਰ These are also called conjunct consonants. A sentence illustrating Gurmukhi script is given below. Terminology used throughout the paper is: TT denotes phonetic transcription, G denotes gloss, and E denotes English sentence.

11 A Punjabi to Hindi Machine Transliteration System 83 ਪ ਜ ਬ ਮ ਰ ਮ ਬ ਲ ਹ TT: pañjābī mērī māṃ bōlī hai. G: Punjabi my mother tongue is. E: Punjabi is my mother tongue. The Nāgarī (lit. 'of the city') or Devanāgarī ('divine Nagari') was originally developed to write Sanskrit but was later adapted to write many other languages, including Hindi. Devanagari has 65 consonants, 18 full vowel letters, 17 vowel symbols, and 2 symbols for nasal sounds. Hindi uses only 34 consonantal syllables, 11 vowel letters (independent vowels), 9 vowel symbols (dependent vowels), and 2 symbols for nasal sounds ( The Devanagari Script Block, 2008). Similar to Gurmukhi, the form of each syllable is often dependent on whether it occurs in combination with other letters, e.g., a vowel following a consonant changes its form, behaving as a diacritic modifier to the consonant: क [ka] + आ [ā] = क [kā] When two consonants occur together, the combination results in a conjunct character. The first sign receives the diacritic known as a halant, to show that the consonant has its inherent vowel silenced, also known as the dead consonant form. क [ka] + र [ra] = कर Unlike Devanagari, there are only a few conjunct consonants in Gurmukhi. In Devanagari, if a consonant is followed by two vowels, then the first vowel is written in the diacritic form, whereas the second is written in full. This feature is not present in Gurmukhi. Vowels may also be nasalized, which is indicated by a anusvara or chandrabindu written over the head stroke (horizontal line) of the vowel or the consonant it is attached to. ह [hũ] nasalized In cases where part of the vowel is written above the head stroke, only the dot is used to indicate nasalization. Although chandrabindu is not a part of Standard Hindi, in practice, both are used interchangeably in Hindi. The sentence illustrating Hindi is given below: प ज ब म र म त भ ष ह paṁjābī merī mātṛbhāṣā hai G: Punjabi my mother tongue is. E: Punjabi is my mother tongue. Except for minor differences, most of the letters are same in both of the scripts. There are three syllables of consonant clusters in Hindi that are written as a single atomic grapheme, i.e., तर, ज ञ, क ष but no such letters or consonant clusters are available in Gurmukhi. Punctuation in

12 84 Gurpreet Singh Josan and Gurpreet Singh Lehal Punjabi is similar to Hindi. Table 1, adapted from (Goyal & Lehal, 2009), shows the similarities and dissimilarities among alphabets in both scripts. As we can see, most of the letters have one to one correspondence, so this table also forms the base for direct mapping. Table 1. Alphabet of Gurmukhi and Devanagari scripts Gur mu khi Dev ana gari Gur mu khi Dev ana gari Gur mu khi Dev ana gari Gur mu khi Dev ana gari Gur mu khi Dev ana gari Gur mu khi Dev ana gari Gur mu khi Dev ana gari ੳ - ਕ क ਟ ट ਨ ऩ ਲ ल ਗ਼ ग़ ਅ अ ਖ ख ਠ ठ ਪ प ਲ਼ ळ ਜ਼ ज़ ੲ - ਗ ग ਡ ड ਫ फ - ऴ ੜ ड़ ਆ आ ਘ घ ਢ ढ ਬ ब ਵ व ੜਹ ढ़ ਇ इ ਙ ङ ਣ ण ਭ भ ਸ਼ श ਫ਼ फ़ ਈ ई ਚ च ਤ त ਮ म - ष ਯ य़ ਉ उ ਛ छ ਥ थ ਯ य ਸ स ਤ ਊ ऊ ਜ ज ਦ द ਰ र ਹ ह - ਏ ए ਝ झ ਧ ध ਰ ऱ ਕ क़ ਐ ऐ - ऋ ਞ ञ ਨ न - र ਖ਼ ख़ ਓ ओ - ॠ - क ਔ औ - ॐ ੴ - 4. Approach to Transliteration from Gurmukhi to Devanagari 4.1 Letter to Letter Mapping Both Punjabi and Hindi are phonetic languages, and their scripts represent the phonetic repository of their respective languages. These phonetic sounds are used to determine the relations between the alphabets of the two scripts. On the basis of this idea, character mappings are determined. The development of the Unicode system for representing alphabets eases the problem to some extent. With this system, every letter can be uniquely mapped to the corresponding letter, as shown in Table 1. Taking into account the similarity of both of the scripts, letter to letter mapping is the obvious choice for the baseline computation. Letters are mapped using Table 1. This system is used as the baseline for improvements by the rule based

13 A Punjabi to Hindi Machine Transliteration System 85 and Soundex based approaches. For analysis and comparison purposes, both scripts are subdivided into different groups on the basis of types of letters, e.g., consonants, vowel symbols, vowel sounds, and other symbols Vowel Mapping Punjabi contains 10 vowel symbols. Hindi vowels have one to one correspondence with Punjabi vowel symbols. There are nine dependent vowel sounds available in Punjabi as,,,,,,,, and. Corresponding vowel sounds in Hindi are,,,,,,,, and. When a vowel sound comes at the start of a word or is independent of any consonant in the middle or the end of the word, independent vowels are used. The mapping is shown in Table 2. Table 2. Vowel Mapping Independent vowels Dependent vowels Gurmukhi Devanagari Gurmukhi Devanagari ਅ अ, ਆ आ, ਇ इ, ਈ ई, ਉ उ, ਊ ऊ, ਏ ए, ਐ ऐ, ਓ ओ ਔ औ Besides these vowels, Punjabi has two more symbols, ੳ and ੲ. No symbol is present in Hindi for these two symbols. The nearest phonetic letters are उ and इ, respectively. These are used in the letter to letter mapping scheme.

14 86 Gurpreet Singh Josan and Gurpreet Singh Lehal Consonant Mapping Consonant mapping is shown in Table 1, Columns No letter in Punjabi is present for the Hindi letters ऴ, ष, and. This means these letters can never be mapped in a letter to letter based approach. Similar is the case for some double-sound-producing syllables, like क ष, ज ञ, and शर. Syllables like ज ञ and शर can be mapped by a rule based approach while others like ष, क ष, ऴ, and can be handled by the Soundex based approach discussed in the next sections Conjunct Consonants Mapping There are three conjunct consonants in Gurmukhi also called PAIREEN, ( Raaraa ), and ( Vaavaa ), which are shown in Table 3. Table 3. Sub Joins of Gurmukhi Conjunct Consonants Punjabi Hindi English ( Haahaa ), ਬ ਲਹ ੜ [bulharh] ब ल हड़ A person with large lips ਪਰ ਤਮ [pritam] पर तम A person name (Pritam) ਸਵ ਰਗ [svarag] स वरग Heaven The usage of PAIREEN Haahaa is quite frequent in Punjabi, but the other two are very rare in usage and are used in Sanskrit loan words. In Punjabi, they are represented by the halant character before the consonant, which indicates that the inherent vowel is omitted. A similar halant character is also present in Hindi, which can replace the halant character in Punjabi. It may be noted that the actual sequence of letters is the same in both languages but the visual sequence of letters differs (Table 4). PAIREEN Haahaa in Punjabi is replaced with their full consonant counterparts, while the full consonant preceding them in Punjabi is shown in half form in Hindi. Similar is the case with PAIREEN Vaavaa. On the other hand, PAIREEN Raaraa takes the position below the previous consonant in Punjabi as well as in Hindi. Table 4. Actual Sequence of letters ਬ ਲ ਹ ੜ = ਬ ਲਹ ੜ ब ल ह ड़ = ब ल हड़ Other Symbols Punctuation marks and digits are the same in both scripts. A special character called as visarga ( ) is present in Hindi but not in Punjabi. So, it will never be mapped in the letter to letter

15 A Punjabi to Hindi Machine Transliteration System 87 based scheme. Besides this, Gurmukhi has two separate nasal characters, Bindi ( ) and Tippi ( ). Hindi has only one nasal character called anusvara. In the Devanagari script, anusvara is represented with a dot (bindu) above the letter e.g. म. Both nasal characters of Punjabi are mapped to this single nasal character in Hindi. Adhak ( ) is used to duplicate the sound of a consonant in Punjabi. No such character is present in Hindi. Sound duplication is represented by half form consonants in Hindi Problems in Letter to Letter Mapping The total number of letters in Punjabi and Hindi are not same (Table 1). The Punjabi letters ੳ and ੲ have no mapping in Hindi. Similarly, there are letters in Hindi that have no mapping in Punjabi e.g. ऋ. These letters will never be mapped in Punjabi to Hindi transliteration using a direct mapping method. Some letters have more than one representation in Hindi, e.g., ਸ਼ may be mapped to श or ष. Another problem is the use of conjunct consonant forms in Hindi. In Hindi, a syllable may consist of a vowel, a consonant followed by vowel, or a consonant cluster followed by a vowel. The last form i.e., when two or more consonants are used within a word with no intervening vowel sound, is known as a conjunct consonant. Use of conjunct consonants is limited in Punjabi. Only three letters can be used as conjuncts i.e., ਹ, ਰ, and ਵ. Their representation is also unique. It is not a trivial task to find out which combinations of letters in Punjabi will take conjunct consonant form in Hindi. For example, why the word ਨਊ [niū] (new) in Punjabi takes the conjunct consonant form in Hindi न य, is not clear. Also, the mapping of nasal consonants is not clear. Nasal consonants in initial place in a conjunct may be expressed using the anusvara over the previous vowel, rather than as a half-glyph attached to the following consonant. The anusvara is written above the headstroke, at the right-hand end of the preceding character. In the list below, both spellings are correct and equivalent, although anusvara is preferred in the case of the first two: र ग = रङ ग, प ज ब = प ब, हद = हन द, ल ब = लम ब. Anusvara is still applied when previous character has its own vowel sign. If the vowel sign is [aa], the anusvara appears over the [aa], e.g. स स or आ द लन. Similarly, the position of the character र is not specific. It can be used as consonant, subscripted consonant, or can take a position above a consonant or diacritics, and it is typically displayed as a small mark above the right shoulder of the last letter in the syllable. It is not clear how the character ਰ [ra] in Punjabi will map to which form in Hindi, e.g., the three cases where mapping of ਰ is not clear are: from ਤਰਕਸ਼ [tarkash] (arrow-holder) to तरकश, from ਵਰਤ [varat] (fast) to वर त, and from ਬਰਤਨ[bartan] (utensil) to बतर न.

16 88 Gurpreet Singh Josan and Gurpreet Singh Lehal 4.2 Rule Based Approach Character mapping alone is not sufficient for a Punjabi to Hindi machine transliteration system. Quite a reasonable improvement can be achieved by small amount of dependency or contextual rules. These rules are manually crafted with the help of a linguist by observing the problems in the direct mapping system. This section discusses such rules for alleviating some of the problems discussed in the previous section. Ideally, the rule set must contain all of the rules required by the system to perform, but practically it is not possible to construct such a set. About 11 rules are identified that are applicable on the source text and about 8 rules are identified that are applicable on the target text. We tried to cover all the possible cases but still we may miss some of them. Moreover, new rules can be added when identified. The rules go as follows: 1. Letters whose mapping is not available in Hindi : a. ੳ and ੲ are two such letters whose corresponding Hindi letters are not available. They are replaced by their most phonetically equivalent letters, i.e., उ and इ respectively. Formally, the rule goes like this if inpt_str contain ੳ then replace it with उ if inpt_str contain ੲ then replace it with इ b. A character adhak is present in Punjabi and used to show the stress on the next character. No letter in Hindi is present to represent this character. Nevertheless, the purpose is served by placing a half character before the stressed character, e.g., ਸ਼ ਕਰ [śakkar] (jaggery) is transliterated as शक कर [śakkar]. There is an exception for this rule. If the next character of adhak is ਖ [kha] then, instead of placing a half character, a half क[ka] is placed, e.g., ਮ ਖਣ [makkhaṇ] (butter) transliterated to मक खन [makkhana]. Also, if the next character is ਛ [ccha], then the half character is replaced by a half च, e.g., as in ਮ ਛਰ [macchar] (mosquito), which is transliterated to मच छर [macchar]. Similarly, if the next character is ਧ [dha], then the half character is replaced by a half द, e.g., as ਬ ਧ [baddh] is transliterated as ब. Formally, it is If current_char = then If nxt_char= ਖ then map to क else if nxt_char= ਛ then map to च else if nxt_char= ਧ then map to द else map to nxt_char & end if end if

17 A Punjabi to Hindi Machine Transliteration System Bindi and Tippi both are used to produce a nasalized sound in Punjabi. There is only Bindi (anusvara) present in Hindi to serve the same purpose. Tippi can be successfully mapped to Bindi in Hindi. A special case is when the tippi [ṃ] is followed by ਨ [na] or ਦ [da], then it is replaced by a half न, as in ਕ ਨੜ [kannaṛ] (Kannad) क ड़ [kannaṛ]. If, however, it is followed by ਬ [ba], then it is replaced by a half म, as in ਮ ਜ਼ ਬਕ [mōzmbik] (mozambique) म ज़ म बक. Formally, it is If current_char = then If nxt_char= ਨ then map to न else if nxt_char= ਬ then map to म else map to end if end if 3. The presence or absence of the diacrtic mark in the target string is effected by the character ਹ [ha] in Punjabi. When ਹ [ha] in the input string is followed or preceded by the [i] mark, [i] is translitrated differently. In this case, ਹ [ha] is followed by then is ommited from the transliterated text, e.g., ਸ਼ ਹਰ [shahir] (city) is transliterated to शहर. On the other hand, if ਹ [ha] is preceded by, then is mapped to in the transliterated text, e.g., ਸਹਤ [sihat] (health) is transliterated to स हत. Formally, it is If current_char = ਹ then If nxt_char= then Omit in target string else if prev_char= then map to else map to end if end if 4. Another rule deals with transliteration of letters ਅਰ [ar]. If this combination appears at the final position in a word, then, instead of mapping ਅ [a]to अ, this letter is mapped to य., e.g., ਬ ਅਰ [bīar] (bear) is transliterated to ब यर. Formally, it is

18 90 Gurpreet Singh Josan and Gurpreet Singh Lehal If last_two_char = ਅਰ then map ਅ to य else map ਅ to अ end if 5. If ਅ is in the last position, then it is replaced by य as in ਪਰ ਅ [pria] पर य Formally, it is If last_char = ਅ then map ਅ to य 6. If ਈ is in the last position, then it is replaced by य, e.g., in ਵ ਜਪ ਈ [vājapāī] (Vajpai) व जप य Formally, it is If last_char = ਈ then map ਈ to य 7. If ਓ is in the last position or the 2 nd to last position, then it is replaced by य, e.g., in ਗਲ ਲਓ [galīliō] (galilio) गल लय. Formally, it is If last_char = ਓ or second_last_char= ਓ then map ਓ to य 8. A rule for plural forms of Punjabi words calls for changing the vowel sign of the last chararcter. It says that, if the last characters are, then is mapped to in Hindi, e.g., ਕਤ ਬ [kitābāṃ] (books) is transliterated to कत ब [kitābō ]. Although this is a case of translation where the root ਕਤ ਬ [kitāb] is the same in the two languages and gets inflected by the rules in respective languages, this case can be handled at the transliteration stage. Formally, it is If last_two_char = then map to 9. Similarly, if the last positions are occupied by ਆ [āṁ], then ਅ [a] is mapped to य, e.g., ਸ਼ ਸ਼ ਆ [shīshīāṃ] (bottles) should be mapped to श श य rather than श श. It is worth mentioning here that both of the spelling variants are acceptable in general to Hindi speakers but the former is preferred. Formally, it is: If last_two_char = ਆ then map ਅ to य 10. The character ੴ [ik-ōṅkār] is replaced by string एक क र. Formally, If char = ੴ then map ੴ to एक क र 11. The letter ਣ [ṇ] is converted to न if it is a last consonant in a word, e.g., in ਕ ਹਣ [kahiṇā] (to say) the letter ਣ is converted to न when transliterated to produce कहन. If last_consnt = ਣ then map ਣ to न The above rules are applicable for the character sequence found in the source script. Table 5 describes some rules that are applicable to the character sequence found in the target script, i.e. Hindi. We can find a probable character sequence using the letter to letter mapping approach. These rules then can be applied to these probable strings. Every row of Table 5 gives the

19 A Punjabi to Hindi Machine Transliteration System 91 replacement rules for a character combination found in the probable string. For example, for a source script word ਵਉ [viu] (view), the letter to letter mapping approach produces a probable string वउ, which is converted to by applying the rule from the following table. Formally, it can be described as follows: For each chr_seqnce in table If chr_ seqnce present in input string then Replace chr_ seqnce to corresponding seqnce from table Table 5. Replacements in transliterated text Occurrence in hindi token Replacement in hindi token Example " उ" " य " ਵਉ [viu] वउ " ऊ" " य " ਨਊ [niū] नऊ न य " ओ" " य " ਬਓਰ [biōrā] बओर ब य र " आ" " य " ਕ ਰ ਆ [kōrīā] क र आ क रय " अ" " य" ਕ ਰ ਅਨ [kōrīan] क र अन क रयन "इण" "यण" ਰ ਮ ਇਣ [rāmāiṇ] र म इण र म यण " य " " य " ਟ ਕ ਓ [ṭōkīō] ट क य ट क य " आ" " य " ਸਆ ਚਨ[siācin] सआ चन स य चन 4.3 Soundex Based Approach Considerable improvements are noticed over the letter to letter mapping approach when using the rule based approach, but still some problems persist. Although we tried to cover the maximum number of rules, still there are cases for which no rule applies. Letters or syllables not available in Punjabi but present in Hindi, ऋ, क ष,ज ञ,शर,ष, etc. are still unhandled. The syllables having double tones in Hindi, like ऋ is a combination of sound of consonant र /r/ and vowel /i/, are represented in Punjabi with the help of two letters, like the consonant ਰ /r/ and vowel /i/. There is no means to find when these two letters in Punjabi will map to corresponding letters र or to a single syllable ऋ in Hindi, e.g., consider following cases. Case I: ਰਕਸ਼ (riksha) [rickshaw] रक श Case II: ਰਸ਼ (rishi) [cleric] ऋ ष

20 92 Gurpreet Singh Josan and Gurpreet Singh Lehal In Case I, the letters ਰ are mapped to र, whereas, in Case II, they are mapped to ऋ. Also, the rule based approach is deficient in producing the half form of letters and some other letters, as discussed in the previous section. For generating these characters, the Soundex technique is employed. Soundex is a phonetic matching technique. The Soundex algorithm was designed by Odell and Russell (1918) to find spelling variation of names. It represents classes of sounds that can be lumped together. Two names match if they have the same Soundex representation. Back in 1918, Odell and Russell proposed Soundex, the first phonetic encoding for English to be used in the US census. Soundex partitions the set of letters into seven disjoint sets, assuming that the letters in the same set have similar sounds. Each of these sets is given a unique key, except for the set containing the vowels and the letters h, w, and y, which are considered to be silent and are not taken into account during encoding. For example, both realize and realise have been encoded to R-420 in Soundex encoding, which works well for trivial cases, but fails to give same code to words where letters change its pronunciation in different contexts. For example, knight, night, and nite are similar sounding words but Soundex does not give the same code to these words. For Punjabi to Hindi transliteration, the Soundex concept is extended for searching for the correct spelling variant of a given transliteration. The Soundex codes are assigned to the Hindi characters based upon their phonetics i.e., similar sounding characters get the same code, e.g., both श and ष have the same sound, thus, they get the same code, 44. The akharas having single graphemes get the code according to the sounds they produce, e.g., ऋ is a combination of sound of consonant र /r/ and vowel /i/, so it gets the code of these two letters, i.e The code table is shown in Table 6. Codes in this table are selected in such a way that similar sounding letters get the same code. We start with independent vowel sounds whose code ranges from 1 to 5. For consonants, we start from two digit codes, i.e. 11 to 47. Thereafter, we code the dependent vowels and some other miscellaneous letters. Unlike Soundex code, which consists of a letter (the first letter of the string) followed by three numerical digits encoding the remaining consonants, all characters of the Hindi string are coded into digits. This ensures matching in case the first letter does not match. From the transliterations produced by the previously discussed methods, the correct spelling variant is searched using Soundex codes as described in the following lines. Let H be a set of equivalent Hindi words and TH be a transliteration produced by the machine transliteration step. A relevant h from H is selected by comparing phonetic similarity between H and TH. For implementation purposes, we make a unigram table from a corpus of about 1 GB size obtained from the Internet. Unigram contains about 150,000 unique words along with their frequency in the corpus. The codes are generated for each of the word in unigram using Table 6. For the comparison, letters in TH are converted into phonetic code using the mapping table as described in Table 6. Then, this code is looked for in a unigram table. The unigram table may

21 A Punjabi to Hindi Machine Transliteration System 93 produce more than one candidate. The candidate with maximum frequency is selected as the correct variant of the given TH produced by the machine. For example, consider the word ਡਰ ਫਟ [ḍarāphaṭ] (draft) written in Punjabi. The string डर फट is produced by the baseline module. The code is generated for this string, as shown in the following table. ड र फ ट This code is looked for in a unigram database. This database contains two entries against this code, i.e. डर फ ट and डर फ ट with frequency 12 and 8. The string with higher frequency is selected as the correct output for this input. Table 6. The coding scheme for Hindi text अ 1 च 17 न ऑ 1 छ 18 प इ 2 ज 19 फ ई 2 ज़ 19 फ़ उ 3 ज ञ 20 ब ऊ 3 झ 21 भ ए 4 ञ 22 म ऐ 4 ट 23 य ओ 5 ठ 24 र औ 5 ड 25 ऌ क 11 ड़ 26 ल 42 or 53 क़ 11 ढ 27 ळ क ष 12 ढ़ 28 व 43 ॐ 55 ख 13 ण 29 श 44 ऋ 4149 ख़ 13 त 30 ष 44 ॠ 4149 ग 14 थ 31 आ 45 or 4149 ग़ 14 द 32 स 46 शर 4441 घ 15 ध 33 ह 47 ङ 16

22 94 Gurpreet Singh Josan and Gurpreet Singh Lehal 5. The Experiment In this section, we describe the evaluation of our models on the task of Punjabi to Hindi transliteration. 5.1 Data A provocative question for evaluators is the type of test material that should be adopted for evaluation. Balkan (1994) describes three types of test material, viz. test corpora, test suites, and test collections. While test corpora is a collection of naturally occurring texts, a test suite is a collection of artificially constructed inputs, where each input is designed to probe a system's treatment of a specific phenomenon or set of phenomena. A test collection is a set of inputs associated with a corresponding set of expected outputs. Test suites and test corpora are extensively used for evaluating various systems involving NLP, while test collections are rarely used. Users can make their own test suites, and such test suites are helpful for checking the improvement of an MT system. For two competing systems, however, test suites are not considered a good method of assessment. It is difficult to interpret the result that both systems transliterate same percentage of text but fail on different sentences. How to design a test suite, the cost of constructing it, and what sort of sentences/words should go into a test suite are some other issues discussed by Arnold (1995). There are several issues involved in the selection of a set of words for a comprehensive evaluation. For example, the set could be constant, variable, or a mixed one; the collection of words may be domain-specific or generic. It is obvious that there is no guarantee that even the bulkiest sample will include all of the possible words of the source language. Some suggestions as to what should go into a test suite are discussed in Hoard (1991). Some empirical evidence is present in literature to answer the question of how much text to include in a test suite. Although this evidence is not for transliteration system, it can provide insight for this system as well. Elliott et al. (2003) tried to prove the general hypothesis that more text would lead to more reliable scores. Based on an empirical assessment of score variation, the authors estimate that systems could be reliably ranked with test suite of around 14,000 words. Zhang & Vogel (2004) also studied the influence of the amount of test data on the reliability of automatic metrics, focusing on confidence intervals for BLEU and NIST scores. Their results show that BLEU and NIST scores become stable when using around 40% of the data (around 40 documents or 300 sentences), although stability is defined here in terms of the distance between scores of different systems. Estrella et al. (2007) shows that, for human or automatic evaluation, about five documents from the same domain-with 6,000 words-seem sufficient to establish the ranking of the systems and about ten documents are sufficient to obtain reliable scores.

23 A Punjabi to Hindi Machine Transliteration System 95 For this experiment, a benchmark sampling method is used to select the set of words. Unicode encoded text is used in Punjabi as well as Hindi, as it eliminates the need of normalizing the text. If text is in font encoded form, then the ASCII code for a given character may vary. For example, under the AnmolLipi font, the Latin character A would appear as ਅ. Conversely, under the DrChatrikWeb font, it would appear as ੳ. Unicode encoded text provides a unique way of representing a character; hence, it eases the task of mapping letters from different alphabets. We tested on people names, location names, and foreign words in Punjabi. The list was prepared manually by typing the names and terms from telephone directories, census records, etc. and partially obtained from manually searching the Internet. We tried to include the maximum possible variety of words in the set. The words are further categorized as Punjabi names, Hindi names, English names, and other foreign names/words. It is worth mentioning here that quite a number of English words are adapted into the Punjabi and Hindi languages, so we categorized them separately from the foreign words, which are words from other languages. The list contains about 3500 person names, 1500 location names, and 1000 foreign words. 5.2 Evaluation Methodology The following combinations of approaches are tested for Punjabi-Hindi transliteration task. Baseline: As a baseline for our experiments, we used a simple letter to letter based approach that maps Punjabi letters to the most likely letter in Hindi. We call it CASE-I. Baseline followed by Rule Based Approach: Some hand crafted rules are studied for the improvements in the letter to letter based mapping system. This system is called CASE-II. Baseline followed by Soundex Approach: The Soundex approach is applied to the output of the baseline method. This is termed as CASE-III. Baseline followed by Rule Based followed by Soundex Based Approach: Both the rule based and Soundex approaches, consecutively, are applied to the output of the baseline method. This is termed as CASE-IV. Baseline followed by Soundex Based followed by Rule Based Approach: Both the Soundex based and Rule based approach, consecutively, are applied to the output of the baseline method. This is termed as CASE-V. Human: For the purpose of comparison, we allowed an independent human subject (fluent in Punjabi and native speaker of Hindi) to perform the same task. The subject was asked to transliterate the Punjabi words in the test set without any additional context. No additional resources or collaboration were allowed. This output was used as the gold standard for checking the system's performance.

24 96 Gurpreet Singh Josan and Gurpreet Singh Lehal 5.3 Evaluation Metrics Often, more than one transliteration is acceptable for any input string. Using one gold standard may prove to be too rigid for measuring accuracy. On the other hand, including more than one correct transliteration complicates the computation of the evaluation score. Also, there is no one exact transliteration of a foreign word, so a soft standard is required to indicate the closeness of output to the gold standard. Levenshtein distance is used for this purpose. The metrics are described as follows: Word Accuracy Rate: It is defined as percentage of correct transliteration from the total generated transliterations by the system. Average Levenshtein Distance: This is the average of Levenshtein distances between the transliterated word and reference word. Levenshtein distance is a measure of similarity of two strings. The distance is the number of deletions, insertions, or substitutions required to transform the source string to the target string, e.g., transforming from the source string इख तआर, which results from the direct transliteration of ਇਖ ਤਆਰ [ikhtiār], the correct target string इखत य र can be produced by deleting from त, adding halant ( ) to त, and replacing अ with य. Thus, the Levenshtein distance is 3. While Levenshtein distance captures the performance of a system at character level, word accuracy rate is used to capture the performance at word level. A word may have non-zero Levenshtein distance but still may be accurate. Lower Levenshtein distance value implies that, although the result is not the same as that of the gold standard, it is still intelligible. 6. Evaluation Results The experiment was performed on the prepared list. The results of this experiment follow. Table 7. Word Accuracy Rate and Avg Lev Dist. of Fivr Cases CASE-I CASE-II CASE-III CASE-IV CASE-V WAR 73.13% 78.88% 86.72% 92.65% 87.15% ALD The baseline model produced a 73.13% accuracy rate. The rule based enhancement and Soundex based enhancement shows the improvements in performance with Soundex prove to be better than rule based. It is observed that CASE-V, i.e. Soundex followed by rule based, does not provide much improvement. This is because in the literature, the Soundex approach is largely discussed as a spelling correction method. Thus, it performs better on refined text, i.e. the output generated by the rule base, and fails to perform on raw input. Similarly, the average edit distance of different cases is shown in Table 7. The average edit distance of

25 A Punjabi to Hindi Machine Transliteration System 97 CASE-III is increased from CASE-II then decreased with WAR increasing, as evident from Figure 2. This is because the human evaluator has marked some spelling variations correct. That is, for the name ਸ ਰ ਦਰ (Surinder), the three transliterations produced by CASE-I, Case-II, and Case-III are स रदर, स रन दर and स रदर, respectively. The human evaluator marked all of these as correct. Nevertheless, as Levenshtein distance is measured against a gold standard, which is स रदर in this case, CASE-III shows an increase in this parameter despite increased word accuracy. The word accuracy rate for CASE IV, i.e. the Base + rule + Soundex approach is found out to be 92.65% and Avg. Levs. Dist is These figures are the highest among all the cases and verify the suitability of the approach. WAR---> Figure 2. Word Accuracy Rate and Avg Lev Dist. of Five Cases The breakdown of the figures is shown in following tables. CASE-I CASE-II CASE-III Table 8. WAR and ALD for person names, location names, and foreign words in all cases Person Name Location Name Foreign Words WAR (%) ALD WAR (%) ALD WAR (%) ALD CASE-I CASE-II CASE-III CASE-IV CASE-V CASE-IV CASE-V ALD---> WAR ALD

26 98 Gurpreet Singh Josan and Gurpreet Singh Lehal WAR---> CASE-I CASE-II CASE-III CASE- IV CASE-V Person Name Location Name Foreign Words Figure 3. WAR for person names, location names, and foreign words in all cases Levsh. Dist. ----> Person Name Location Name Foreign Words 0 CASE-I CASE-II CASE-III CASE- IV CASE-V Figure 4. ALD for person names, location names, and foreign words in all cases The word accuracy for words whose origin is also Punjabi is 86.6 in the baseline model and shows the similar trend in other cases, as shown in Table 9. The improvement in all cases is registered with maximum for Hindi and English languages. Table 9. Word Accuracy rate and Avg. Lev Dist according to origin of source language of input word Punjabi Word Hindi Word English Word Foreign Word WAR ALD WAR ALD WAR ALD WAR ALD CASE-I CASE-II CASE-III CASE-IV The accuracy in the baseline model for Hindi and English is quite low because the baseline model cannot capture the half form representation of letters. As there is no half form in

27 A Punjabi to Hindi Machine Transliteration System 99 Punjabi, a Hindi name, when spelled in Punjabi, uses the full form of the character instead of its half form. For example, the name इन दर (Inder) when spelled in Punjabi will look like ਇ ਦਰ [indar]. Here, the half form of न and र in the Hindi name are represented by (tippi-a character for nasal sound) and ਰ, respectively. When this form is transliterated by the baseline model, it will produce इ दर, which is incorrect. Similar is the case with English and other foreign words. So, due to the character gap in Punjabi and other languages, the word accuracy rate for the baseline is low. It is also interesting to note that when words that are originally from Hindi and used in Punjabi are transliterated back to Hindi, the accuracy rate is lower than other types of words (for Cases III and IV). The reasons are twofold. First, Hindi words written in Punjabi are an approximate transliteration of the original Hindi word in Punjabi. Depending upon the perception of the transliterator, a Hindi word may have more than one representation in Punjabi, e.g., the Hindi name आच य र (Acharya) can be represented in Punjabi as ਆਚ ਰਆ, ਆਚ ਰ ਆ, ਅਚ ਰ ਆ, and ਅਚ ਰਆ. Only the first representation, when again transliterated back into Hindi, converts to the correct representation. Other representations are converted into such strings by baseline and rule based modules that they have no match in the unigram table, making the Soundex module inefficient. Normalization of spellings at the source by a similar Soundex approach may improve the results. Table 10. Breakdown of Word Accuracy rate and Avg. Lev Dist for person names, location names, and other words according to origin of source language of input word Punjabi Words Hindi Words English Words Foreign Words WAR ALD WAR ALD WAR ALD WAR ALD Person Name CASE-I Location Name Other Words Person Name CASE-II Location Name Other Words Person Name CASE-III Location Name Other Words CASE-IV Person Name

28 100 Gurpreet Singh Josan and Gurpreet Singh Lehal CASE-V Location Name Other Words Person Name Location Name Other Words The system also benefits from the spelling correction capability of the Soundex method. It can improve the results by selecting the correct output for a given wrong input, e.g., for the input ਅਨ ਪਮ (correct is ਅਨ ਪਮ [anupam]), the Soundex method selects the correct variation अन पम instead of अन पम. On the other hand, at times it selects the wrong target word, e.g., for input ਕ ਟ [kaiṇṭ] (cantt) the correct output is क ट but the Soundex method selects क ट as output. Another problem with the Soundex based approach is that it is dependent upon the unigram table. Although this table is generated from a large amount of data, still it is not exhaustive. New terms and names are coined every day, and these need to be included in the unigram table regularly. If the data is not present in this table, then the system fails. Such cases, however, are few and overall accuracy of the system improves with the combined approach of baseline + rule based + Soundex based approach. 7. Conclusion The results show that following the rule based followed by Soundex approach produces the best results and the words can be transliterated with considerable accuracy. A fully accurate transliteration system is not possible due to the inherent problems, missing corresponding letters in two scripts, conjunct form, etc. Although it is possible to transliterate across the scripts preserving the basic sounds of the source language, there will be some variations because the word in the source script may be pronounced somewhat differently in the target script, as per local conventions in each region. This results in more time required by the users to interpret the word in the context of original language from which the word comes. The transliterated text will be correctly understood only if the reader has knowledge of the conventions used for representing sounds in the source script. The aim of the module at least is to present a nearest possible transliterated text to the user instead of going blank to such words. This module is successfully used for transliterating proper nouns in Punjabi to Hindi machine translation system and in the Cross Language Information Retrieval system PunjabiKhoj, which is a search engine for Punjabi language. Although the system performs well, still there is room for improvement. Transcription rules based upon IPA and statistical models will be considered in the future. Testing of the

29 A Punjabi to Hindi Machine Transliteration System 101 system on precision, recall, and F-score is on our future agenda. Reference Al-Onaizan, Y., & Knight, K. (2002). Translating named entities using monolingual and bilingual resources. Proceedings of the 40th ACL, Philadelphia, Arnold, D., Balkan, L., Humphreys, R. Lee, Meijer, S., & Sadler, L. (1995). Machine Translation: An Introductory Guide. NCC Blackwell, Manchester, Oxford. Balkan, L. (1998). Test suites: some issues in their use and design. In proceedings of the International conference on Machine translation: ten years on held at Cranfield University, England, November 1994 (Cranfield University Press, 1998). Bhatia, T. K. (1993). Punjabi: A Cognitive-descriptive Grammar. Descriptive Grammars, Routledge, London. Chopde, A. (2001). Printing Transliterated Indian Language Documents, ITRANS, from website (Accessed on 22 Aug 2006). Elliott, D., Hartley, A., & Atwell, E. (2003). Rationale for a multilingual aligned corpus for machine translation evaluation. In International Conference on Corpus Linguistics (CL2003), Lancanster, UK. Estrella, P., Hamon, O., & Popescu-Belis, A. (2007). How much data is needed for reliable MT evaluation? Using bootstrapping to study human and automatic metrics. MT Summit XI, September 2007, Copenhagen, Denmark. Proceedings, Goyal, V., & Lehal, G. S. (2009). Hindi-Punjabi Machine Transliteration System (For Machine Translation System). George Ronchi Foundation Journal, Italy, 64(1), Hoard, J. (1991). Preliminaries to the Development of Evaluation Metrics for Natural Language Semantic and Pragmatic Analysis Systems. in Neal, J.G. and Walter, S.M. (eds.), (1991): Natural Language Processing Systems Evaluation Workshop, Report RLTR , Rome Laboratory. Jung, S. Y., Hong, S. L. & Paek, E. (2000). An English to Korean Transliteration Model of Extended Markov Window. Proceedings of COLING KalyanaSundram, K. Mylai Tamil Font for preparation of Texts in Tamil Script on Computers, from website Kang, B.J. & Choi, K.-S. (2000). Automatic Transliteration and Back-transliteration by Decision Tree Learning. Proceedings of the 2nd International Conference on Language Resources and Evaluation, Athens, Greece. Kanneganti, R. & Kishore, A. Rice Inverse Transliterator (RIT) Software, from website Knight, K. & Graehl, J. (1998). Machine Transliteration. Computational Linguistics, 24(4). Lee, C. & Chang, J. S. (2003). Acquisition of English-Chinese Transliteration Word Pairs from Parallel-Aligned Texts using a Statistical Machine Translation Model. Proceedings

30 102 Gurpreet Singh Josan and Gurpreet Singh Lehal of HLT-NAACL Workshop: Building and Using parallel Texts Data Driven Machine Translation and Beyond, 2003, Edmonton, Malik, M.G.A. (2006). Punjabi Machine transliteration. Proceedings of the 21st International Conference on Computational Linguistics and the 44th annual meeting of the ACL, Meng, H. M., Lo, W., Chen, B., & Tang, K. (2001). Generate Phonetic Cognates to Handle Name Entities in English-Chinese cross-language spoken document retrieval. Proceedings of ASRU, Odell, M.K., & Russel, R.C. (1973). US Patent , 1918, cited in Knuth (1973) Sorting and Searching: The art of computer programming Volume 3, Addison Wesley, reading, MA. Oh, J.H., & Choi, K.S. (2002). An English-Korean transliteration model using pronunciation and contextual rules. Proceedings of the 19th international conference on Computational linguistics, 1, 1-7. Saini, T. S., & Lehal, G. S. (2008). Shahmukhi to Gurmukhi Transliteration System: A Corpus based Approach. Research in Computing Science (Mexico), 33, Srinivasan. (1995). ADHAWIN, in Transliteration schemes from website The Devanagari Script Block. (2008). from website accessed on 7 Feb Transliteration Principles. (2008). from accessed on 7 Feb Virga, P., & Khudanpur, S. (2003). Transliteration of proper names in cross-language applications. In Proceedings of the 26th annual international ACM SIGIR conference on Research and development in information retrieval, Viswanadha, R. (2002). Transliteration of Tamil and Other Indic Scripts. Tamil Internet 2002, California, USA. Wan, S., & Verspoor, C. M. (1998). Automatic English-Chinese name transliteration for development of multilingual resources. Proceedings of COLING-ACL 98. Writing Systems. (2010). form website Zhang, Y., & Vogel, S. (2004). Measuring Confidence Intervals for the Machine Translation Evaluation Metrics. In International Conference on Theoretical and Methodological Issues in Machine Translation (TMI 2004). Baltimore, MD.

31 Computational Linguistics and Chinese Language Processing Vol. 15, No. 2, June 2010, pp The Association for Computational Linguistics and Chinese Language Processing A Posteriori Individual Word Language Models for Vietnamese Language Le Quan Ha, Tran Thi Thu Van, Hoang Tien Long +, Nguyen Huu Tinh +, Nguyen Ngoc Tham +, and Le Trong Ngoc Abstract It is shown that the enormous improvement in the size of disk storage space in recent years can be used to build individual word-domain statistical language models, one for each significant word of a language that contributes to the context of the text. Each of these word-domain language models is a precise domain model for the relevant significant word; when combined appropriately, they provide a highly specific domain language model for the language following a cache, even a short cache. Our individual word probability and frequency models have been constructed and tested in the Vietnamese and English languages. For English, we employed the Wall Street Journal corpus of 40 million English word tokens; for Vietnamese, we used the QUB corpus of 6.5 million tokens. Our testing methods used a priori and a posteriori approaches. Finally, we explain adjustment of a previously exaggerated prediction of the potential power of a posteriori models. Accurate improvements in perplexity for 14 kinds of individual word language models have been obtained in tests, (i) between 33.9% and 53.34% for Vietnamese and (ii) between 30.78% and 44.5% for English, over a baseline global tri-gram weighted average model. For both languages, the best a posteriori model is the a posteriori weighted frequency model of 44.5% English perplexity improvement and 53.34% Vietnamese perplexity improvement. In addition, five Vietnamese a posteriori models were tested to obtain from 9.9% to 16.8% word-error-rate (WER) reduction over a Katz trigram model by the same Vietnamese speech decoder. Faculty of Information Technology, Hochiminh City University of Industry, Ministry of Industry and Trade, 12 Nguyen Van Bao, Ward 4, Go Vap District, Hochiminh City, Vietnam lequanha@ hui.edu.vn; NLp.Sr@ Shaw.ca; letrongngoc@ hui.edu.vn + Faculty of Information Technology, Nong Lam University of Hochiminh City, Block 6, Linh Trung Ward, Thu Duc District, Hochiminh City, Vietnam { , , }@ st.hcmuaf.edu.vn

32 104 Le Quan Ha et al. Keywords: A Posteriori, Stop Words, Individual Word Language Models, Frequency Models. 1. Introduction A human is able to work out the precise domain of a spoken sentence after hearing only a few words. The clear identification of this domain makes it possible for a human to anticipate the following words and combination of words, thus, recognizing speech even in a very noisy environment. This ability to anticipate still cannot be replicated by statistical language models. In this paper, we suggest one way that significant improvement in language modeling performance can be achieved by building domain models for significant words in a language. The word-domain language model extends the idea of cache models (Kuhn & De Mori, 1990) and trigger models (Lau, Rosenfeld, & Roukos, 1993) by triggering a separate n-gram language model for each significant word in a cache and combining them to produce a combined model. The word-domain models are built by collecting a training corpus for each significant word. This is done by amalgamating the text fragments where the word appears in a large global training corpus. In this paper, the text fragments are the sentences containing the significant words. Sicilia-Garcia, Ming, and Smith (2001, 2002) have shown that larger fragments are not needed. We define a significant word as any word that significantly contributes to the context of the text or any word that is (i) not a stop word, i.e. not an article, conjunction, or preposition; (ii) not among the most frequently used words in the language, such as will ; and (iii) not a common adverb or adjective, now, very, some, etc. All other words are considered significant, and a corpus is built for each. A statistical language model is then calculated from this corpus, i.e. from all of the sentences containing the word. Therefore, the model should be able to represent the domain of that word. This approach entails a very large number of individual word language models being created, which requires a correspondingly large amount of disk storage; previous experiments by Sicilia-Garcia, Ming, and Smith (2005) were done on twenty thousand individual word language models, which occupied approximately 180 GigaBytes. Thus, this tactic is feasible only given the relatively recent increase in affordable hard disk space. These word models gradually developed from the PhD work of Sicilia-Garcia Almost at the same time as her research, similar research work was done by Blei, Ng, and Jordan (2003); they originally started from the ideas of Hofmann, Puzicha, and Buhmann (1998) and from Hofmann (1999). The remaining sections are organized in the following way. First, we discuss the weighted average model our developed approach lies upon. Then, we discuss both the probability models - including linear interpolation and the exponential decay model - and the

33 A Posteriori Individual Word Language Models for Vietnamese Language 105 weighted models, such as the weighted probability model, weighted exponential model, and linear interpolation exponential model with weights. Then, we discuss their corresponding frequency models and we discuss a priori and a posteriori testing methods. Following this, the corpus, experiments, and results are shown for the probability models, for the frequency models, and for a posteriori models. Finally, we provide some conclusions. 2. The Language Models Experiments had shown that we needed to combine the global language model with the individual word-domain models in order to obtain good results. (This may be due to the limited size of the global corpus in our tests, which was 40 million tokens.) So, we first built a language model for the whole global corpus. Frequencies of words and phrases derived from the corpus and the conditional probability of a word given a sequence of preceding words were calculated. The conditional probabilities were approximated by the maximum likelihoods: PML where ( 1 n ) i 1 ( wi w1 ) i ( ) i 1 ( w1 ) (... i i) f ( w... w ) f w = 1 = f w1 w 1w (1) f 1 i 1 f w is the frequency of the phrase w1 = w1... wn 1wn in the text. These probabilities were smoothed by one of the well-known methods, such as Turing-Good estimation (Good, 1953) or the Katz back-off method (Katz, 1987). Although any of these could be used in our experiment to demonstrate the principle of our multiple word-domain models, it was convenient to use the empirical weighted average (WA) linear interpolation n-gram model (O Boyle, Owens & Smith, 1994) because of its simplicity. It gives results comparable to the Katz back-off method but is much quicker to use. The weighted average probability of a word w given the preceding words w 1 w m-1 w m is: P WA m ( w w1 ) m m ( ) + μ ( + ) μ P w P w w = 0 ML i= 1 i ML m 1 i m μi i= 0 where the weighted functions (in the simplest case) are given by m i μ0 Ln( T ) and μ i Ln( f ( wm+ 1 i) ).2 m where T is the number of tokens in the corpus and ( m 1 i) = = (3) f w + is the frequency of the sentence w m+1-i w m in the text. The unigram maximum likelihood probability of a word is: PML ( w) ( w) f = (4) T n (2)

34 106 Le Quan Ha et al. The language model defined by Equations (2) and (4) is called the global language model when trained on the global corpus. The creation of a language model for each significant word is formed in the same manner as the global language model. 3. Probability Models We need to combine the probabilities obtained from each word-domain language model and from the global language model in order to obtain a combined probability for a word, given a sequence of words. One simple way to do this is a mathematical combination of the global language model and the word language models in a linear interpolated expression as: where m n n n ( 1 ) λg G ( 1 ) λi i( 1 ) i= 1 m G + λi = 1 and G ( 1 n ) P w w = P w w + P w w (5) λ 1 P w w is the conditional probability of the word w following a phrase w 1...w n-1,w n in the global language model, P i is the conditional probability in the word language model for the significant word w i, λ i is the correspondent weight, and m is the number of word models that are included. Ideally, the λ i parameters would be optimized using a held-out training corpus; however, this is not practical as we do not know which combination of words w i will arise in the cache. So, a simpler approach is needed. 3.1 Linear Interpolation A simple method of choosing the λ-values is to give the same weight to all of the word language models but a different weight to the global language model and to put a restriction on the number of word language models to be included. This weighted model is defined as ( 1 ) m n n λ n P( w w1 ) = λ. PG ( w w1 ) + Pi( w w1 ) (6) m i= 1 and λ and m are parameters that are chosen to optimize the model. 3.2 Exponential Decay Model A method was developed based on an exponential decay of the word model probabilities with distance since a word appearing several words before the target word will generally be less relevant than more recent words. Given a sequence of Vietnamese words, for example, (meaning I had friendly relations with HUI ) in Table 1, where 5, 4, 3, 2, 1 represent the distance of the word from the target word HUI. The words tri (relation) and giao (friendly) are significant words for which we have individual word language models.

35 A Posteriori Individual Word Language Models for Vietnamese Language 107 Table 1. An explanation of distance of words. tri giao This model for the word w, where w represents the significant word HUI, is as follows: n ( 1 ) P w w n n n ( 1 ) + ( 1 ).exp( 3 / ) + ( 1 ).exp( 2 / ) PG ww Ptri ww d Pgiao ww d = 1+ exp 3/ + exp 2/ where G ( 1 n ) in the global language model and tri ( 1 n ) ( d) ( d) P w w is the conditional probability of the word w following a phrase w 1,w 2 w n P w w is the conditional probability of the word w following a phrase w 1 w n-1 w n in the word language model for the significant word tri. The same definition applies for the word model giao. d is the exponential decay distance with d=5, 10, 15, etc. A cache or cut-off is introduced in the model if l cache => replace exp(-l/d) by 0 where l is the distance from the significant word to the target word. 3.3 Weighted Models In the two methods above, the weights for the word language models were independent of the size of the word training corpora or the global training corpus. So, we introduced new weights to these models that depend on the size of the training corpora. These weights are functions of the size of the word training corpora, i.e. the number of tokens of the training corpora T i. Examples of the weights can be seen in Table 2. Table 2. Some of the weights in weighted models. Weights Ln(1+LnT i ) Sqrt(LnT i ) LnT i Sqrt(T i ) T i /LnT i T i (7) T i LnT i

36 108 Le Quan Ha et al. An obvious weight to use is a log function to match the information theory, but other weights were also tried. All of these weights are studied in order of weight difference as T i increases, from the largest difference to the smallest. Their relationship can be seen in Figure 1. log(weight) TiLnTi Ti Ti/LnTi Sqrt(Ti) LnTi Sqrt(LnTi) Ln(1+LnTi) log(ti) Figure 1. Value of the weights used in these models Weighted Probability Model The weighted probability model is based on the idea that the weight given to a word language model should depend on the size of the training corpora. It is described as follows: n ( 1 ) P w w m n n ( 1 ) + β ( 1 ) β. P ww. P ww = G G i= 1 i i m βg + βi i= 1 where β G is the weight for the global language model and β i is the weight for the word model for the word w i. We give more weight to those word models with a small training corpus as they represent models for the less frequent words, which have the most information. The weights used are functions of the size of the word training corpora in Table 2, that is, of the number of tokens of the training corpora T i Weighted Exponential Model The weighted exponential language model is a combination of the weighted probability model and the exponential decay model. Each language model has two functions: one is the exponential decay, in terms of the distance from the significant word, and the second function is a weight, depending on the size of the word training corpora. We define this model as: (8)

37 A Posteriori Individual Word Language Models for Vietnamese Language 109 n ( 1 ) P w w β = m n xi n ( 1 ) + β ( 1 ) G. PG w w i.exp Pi w w i= 1 d m x.exp i βg + βi i= 1 d where x i is the distance from the word w i model. (9) Linear Interpolation Exponential Model with Weights Finally, we decided to try another model that is based on a combination of all previous methods. Based on the idea that the global language model could be weighted in a different way from the word language models, the probability of a word given the previous words is: n ( 1 ) m x.exp i βi Pi w w n n i 1 d = P( w w1 ) = λpg ( w w1 ) + ( 1 λ ) m x β.exp i i i= 1 d This is equivalent to combining all of the methods seen previously into one model, which we call the linear interpolation exponential model with weights. 4. Frequency Models Instead of combining probabilities to obtain a dynamic language model, it is also possible to combine frequencies before calculating probabilities, i.e. a revised maximum likelihood. To do this, replace Equation (1) with: m i i λg fg( w1) + λi fi( w1) i 1 i= 1 PML ( wi w1 ) = m i 1 i 1 λg fg( w1 ) + λi fi( w1 ) i= 1 This can then be combined using the WA model in Equation (2). This simple method is automatically normalized, and it is easy to implement and fast to execute. The choice of λ is still critical but cannot be optimized from a held out corpus. For the frequency model, we also combine the frequencies using the same methods that are used for probabilities. 4.1 Linear Interpolation Frequency Model The linear interpolation model applied to the frequency is defined as: ( ) ( ) λ ( ) ( 1 λ ) ( ) m n n n ffm w1 = ffm w1w2... wn =. fg w1 + fi w1 if m> 0 m i= 1 G 1 n (10) (11) (12) where f ( w ) is the frequency of the phrase w, w w from the global model, ( ) 1 2 n fi w 1 n is the frequency of the phrase w 1, w 2 w n from the word w i model, and m and λ>0 are chosen

38 110 Le Quan Ha et al. parameters. 4.2 Exponential Decay Frequency Model A method was used based on an exponential decay of the frequencies: n n ( 1 ) ( ) ( 1 ) m xi n = = + exp ( 1 ) ffm w ffm w w wn fg w fi w i= 1 d (13) 4.3 Weighted Models We introduce new weights to these models; they are functions shown in Table Weighted Frequency Model The weighted frequency model is a mathematical combination of the word language models with the global language models. The frequency for this model is defined as follows: n n ( 1 ) ( ) ( 1 ) m n β β. ( 1 ) ffm w ffm w w wn G fg w i fi w i= 1 = = + (14) Weighted Exponential Decay Frequency Model As can be seen from the previous section, the frequencies are weighted depending on the size of the training corpora only. In this model, the exponential decaying factor is added to these weights. The new weighted exponential frequency becomes: n n ( 1 ) ( ) ( 1 ) m xi n = = β + β.exp ( 1 ) ffm w ffm w w wn G fg w i fi w i= 1 d (15) 5. Testing Methods Perplexity is a well known measure of the performance of a language model (Jelinek, Mercer, n & Bahl, 1983). We calculate the perplexity of each sentence, w 1, by n n 1 1 = n 1 (16) Pw ( ) Pw ( ) Pw ( w)... Pw ( w ) and the perplexity of a sentence by 1 m n PP( w1 ) = exp Ln( P( wi w1w 2... wi 1) ) (17) m i= 1 where m is the number of words in the sentence. There are two methods of calculating the constituent probabilities on the right hand side of Equation (17) using the word domain language models, one a priori by Sicilia-Garcia et al. (2001, 2002) and the second a posteriori, which was first tried by Sicilia-Garcia et al. (2005.)

39 A Posteriori Individual Word Language Models for Vietnamese Language A Priori Method In the a priori method, we use the global language model and (possibly) individual word models from earlier sentences (i.e. from the cache) at the beginning of the sentence because we do not know which significant words are going to appear in the sentence. We then add in a word language model for each significant word after it appears in the sentence. Thus, in the sentence, The cat sat on the mat, neglecting previous sentences, the first two words are modeled using the global language model, the probability P(sat the cat) is calculated using the global model combined with the word model for cat, and the last three words are modeled using the global model combined with the word models for cat and sat. 5.2 A Posteriori Method This, however, is not the only way in which models are tested. For example, in domain language models, researchers extract a whole sentence, paragraph, or document from the test file, find all of the significant words within it, and use all of these words to perform an optimization of the possible domains to find the domain or combination of domains to minimize the perplexity (Seymore, Chen, & Rosenfeld, 1998; Donnelly, 1998; Iyer & Ostendorf, 1999; Sicilia-Garcia et al., 2005.) To make comparisons with these other domain methods, Sicilia-Garcia et al. (2005) also tried this a posteriori method to calculate perplexity for word-domain models. To do this, they extracted all of the significant words in a sentence and built a language model based on the global model and the word domain models for the significant words. This was then used to calculate the perplexity of the whole sentence. In the example above, The cat sat on the mat, the perplexity was calculated using the global model combined with the word domain models for the three words cat, sat, and mat for the whole sentence calculation. Using this approach, they obtained 68%-69% improvement in perplexity when using the a posteriori weighted probability model and the a posteriori weighted frequency model. They accepted that this 69% improvement exaggerated the performance but showed the potential power of these a posteriori models. Upon further analysis of this, however, we have discovered a slight flaw in the calculation. This can be demonstrated by again considering the example above. They calculated the a posteriori probability P(sat the cat) by employing the global model, the word cat model, the word sat model, and the word mat model. As the unigram sat, as well as its n-grams, obviously occur themselves in every sentence of its own target word sat model, this yields an unnaturally high probability P sat (sat the cat). Hence, the flaw: if we replace the word sat by any significant words xxx in the test, the probability P xxx (xxx the cat) will also obtain unnaturally high values in its own word xxx model. In this work, we propose and test a corrected method for calculating the a posteriori probability P(sat the cat), which uses only the global model and the word models for cat

40 112 Le Quan Ha et al. and mat, excluding the target word sat model. Furthermore, we have tested three different ways of combining word models in our calculation for the linear interpolation probability, the exponential decay probability, and the weighted probability models in the whole sentence test: i. applying the word models appearing from the beginning until the end of each sentence, excluding the target word model ii. iii. applying the word models within the phrase history first then the word models that appear later in each sentence, and ignoring the order of appearance of significant words in each sentence, either existing in the phrase history or occurring later, but applying those significant word models that are located from nearer to farther distances, relative to the target word. We find out that the (iii) calculation provides 1%-3% better perplexity improvements than (i) and (ii) calculations for all three models. This means that nearer word models supply more reliable probabilities. Sicilia-Garcia (2002) suggested another whole paragraph calculation with poorer results. 6. Corpus The methods described above were compared in some Vietnamese and English experiments. For English, we used the Wall Street Journal (WSJ) corpus (Paul & Baker, 1992). Previous research by Sicilia-Garcia et al. (2001, 2002, 2005) displayed how the individual word probability models depend on the size of the training corpus for two subsets of the WSJ of approximately 16 million (1988) and 6 million words (1989). The well-known WSJ test file (Paul et al., 1992) contains 584 paragraphs, 1,869 sentences, 34,781 tokens, and 3,677 word types. In this work, we develop these models for the whole combined WSJ corpus of 40 million words. The results reveal a lower perplexity for the larger 40 million word corpus, compared to Sicilia-Garcia et al. For Vietnamese, we employ the syllabic Vietnamese corpus by Ha in 2002 ( orpus.aspx), with a size of 6,564,445 tokens and 31,402 types. This corpus was collected from popular Vietnamese newspapers, such as The Youth Newspaper, Saigon Liberation Newspaper, Vietnamese Culture and Sports, Motion Pictures Magazine, and Vietnam Express, along with traditional and modern novels and short stories. Vietnamese is the national and official language of Vietnam. It is the mother tongue of 86% of Vietnam's population, and of about three million overseas Vietnamese people. It is also spoken as a second language by many ethnic minorities of Vietnam. It is part of the Austro-Asiatic language family, of which it has its most speakers, having several times more speakers than the other Austro-Asiatic languages put together. Some Vietnamese vocabulary

41 A Posteriori Individual Word Language Models for Vietnamese Language 113 has been borrowed from Chinese, and it was formerly written using a Chinese-like writing system, albeit in a modified format and with vernacular pronunciation. As a byproduct of French colonial rule, the language displays some influence from French, and the Vietnamese writing system in use today is an adapted version of the Latin alphabet, with additional diacritics for tones and certain letters. Vietnamese is basically a monosyllabic language with six tones, which gives the language a sing-song effect. A syllable can be repeated with any one of six tones to indicate six different meanings. For example, the syllable ma has six different meanings according to the tone this Vietnamese syllable carries - with level tone, ma means phantom or ghost ; with low tone, means but, which, or who ; with high rising glottalized tone, means code ; with dipping-rising tone, means tomb ; with high rising tone, means cheek ; and with low glottalized tone, means young rice seedling. Due to the semantic impact of tonality, we would like to apply our language models for Vietnamese syllables instead of English words. We also established our Vietnamese test text from the above Vietnamese newspapers, but its content was taken from newspapers of the year 2008, much later than Ha s training text in Therefore, the Vietnamese test text is totally different from Ha s corpus; it includes 33,108 Vietnamese syllable tokens of 2,226 syllable types within 3,321 sentences. 7. Results We will first present the results for each a priori model, starting from the probability models and moving to the frequency models then show our a posteriori results. All perplexity values shown in the tables are accompanied with a percentage, which shows the improvement compared to the global base-line trigram WA model. 7.1 Results for Probability Models We present the results for all a priori probability models, starting from the linear interpolation probability model and going to the linear interpolation exponential decay probability model with weights. It can be seen from the results that the best English and Vietnamese probability model is the weighted exponential probability model, with overall results in Tables 3 and 4. The best performance achieved using each different type of probability models is shown in Table 5 and Table 6. The best model is the weighted exponential probability model, with 34% improvement for English and 37% for Vietnamese, while the linear interpolation exponential model with weights - our combination of all of the models - disappointingly only improves 32% for English and 33.9% for Vietnamese. For these models, the number of individual word models

42 114 Le Quan Ha et al. required in the cache to reach the maximum performance is English words and Vietnamese syllables. So, the individual word-domain language model reduces the size of the cache needed from 500 words, as in other models (Clarkson & Robinson, 1997; Donnelly, Smith, Sicilia-Garcia & Ming, 1999), to less than 30 English words or 64 Vietnamese syllables, which is important for spoken language and is closer to the ability of humans. In Table 5 and Table 6, WM represents the number of word language models that is m in Equations (6), (8), (9), and (10). Table 3. The weighted exponential model (English WSJ.) Sentence/WSJ n-gram Perplexity Improvement (d, Cache, Function) tri-gram % (8, 75, Sqrt(1/LnT i )) 5-gram % (8, 70, LnT i ) 7-gram % (7, 75, LnT i ) 9-gram % (7, 75, LnT i ) Table 4. The weighted exponential model (Vietnamese QUB.) Sentence/QUB n-gram Perplexity Improvement (d, Cache, Function) tri-gram % (13, 100, Sqrt(LnT i )) 5-gram % (13, 100, LnT i ) 7-gram % (13, 100, LnT i ) 9-gram % (13, 100, LnT i ) Table 5. Improvement in perplexity for different probability models, all in sentence contexts (English WSJ.) Models tri-gram 9-gram Best Values Global 0.00% 26.77% Linear Interpolation 11.16% 31.98% λ=0.7, WM=23 Exponential Decay 16.75% 33.72% Decay=6, Cache=70 Weighted Probability 13.90% 32.52% WM=16, Sqrt(T i ) Weighted Exponential 16.78% 33.90% Decay=7, Cache=75, LnT i Linear Interpolation Exponential with Weights 12.91% 32.28% λ=0.6, Decay=13, Cache=65, Ln(1+LnT i )

43 A Posteriori Individual Word Language Models for Vietnamese Language 115 Table 6. Improvement in perplexity for different probability models, all in sentence contexts (Vietnamese QUB.) Models tri-gram 9-gram Best Values Global 0.00% 20.88% Linear Interpolation 18.46% 34.02% λ=0.6, WM=64 Exponential Decay 22.88% 36.54% Decay=10, Cache=100 Weighted Probability 20.45% 34.45% WM=29, Sqrt(T i ) Weighted Exponential 22.98% 36.92% Decay=13, Cache=100, LnT i Linear Interpolation Exponential with Weights 18.40% 33.88% λ=0.6, Decay=57, Cache=100, 1/Ln(1+LnT i )) Our weighted exponential model is a special case of the linear interpolation exponential with weights. Hence, it should not have better performance. Sicilia-Garcia et al. (2001, 2002, 2005) were also disappointed when this combination model of all other models was not good. The reason for this unusual observation is a conflict occurring between the weighted exponential model (that is a combination of exponential decay and weighted probability models) and the linear interpolation model. The linear interpolation model was optimized on the condition that all significant words are equally treated, while the weighted exponential model treats significant words differently from each other. In the linear interpolation model, the global weight is λ = 0.6 and each Vietnamese syllabic model equally optimized at a weight of (1-λ)/m = (1-0.6)/64 = 0.006,25. For the weighted exponential model, however, the global weight is Ln(40M) = 17.5 and a Vietnamese syllabic model with size 10,000 has a weight of Ln(10,000) = Another Vietnamese syllabic model with size 100 has its weight of Ln(100) = On the linear interpolation exponential with weights, when these weights are multiplied or interpolated together, they break the optimization of both the linear interpolation model and the weighted exponential model; the too small individual weight 0.006,25 in the first model and the much larger global weight 17.5 in the latter model are not satisfied. 7.2 Results for Frequency Models We present the results for each a priori frequency model, starting from the linear interpolation frequency model and going to the weighted frequency model. The best English frequency model is the weighted frequency model, and our results for this model are displayed in Table 7. An improvement of 38% has been achieved for the English weighted frequency model. The best Vietnamese one is the exponential decay frequency model with 47.3% perplexity improvement, as shown in Table 9.

44 116 Le Quan Ha et al. Table 7. The weighted frequency model (English WSJ.) Sentence/WSJ n-gram Perplexity Improvement (WM, Function) tri-gram % (29,1/T i *Ln(T i )) 5-gram % (29, 1/T i ) 7-gram % (29, Ln(T i )/T i ) 9-gram % (29, Ln(T i )/T i ) For the English weighted frequency model, it is important to notice that the perplexity result for the 9-gram model is poorer than the one for the 7-gram language model. We think this is because the word language models in many cases are so small that the 9-gram frequencies are usually zero. In order to recognize or understand a rather short phrase as a tri-gram, its meaning largely depends on its context. The other possibility is that the historical word language models need more weight than the large global model of 40 million tokens. If a significant word language model has 1,000 words in its corpus, then the global weight will approximately be 1/(40M*Ln(40M)) = 1.428E-09 while that word model s weight 1/(1,000*Ln(1,000)) = 0.000,144 is much larger in comparison. Nevertheless, for longer phrases, such as 5-grams or 9-grams, the meanings of a very long 9-gram is almost obvious by itself and its meaning is less impacted by significant words surrounding it. Sometimes, people understand a long spoken phrase with 9 continuous words without confusion even though they did not hear the previous significant words. Therefore, for longer English phrases, the global weight should gain importance and increase its value, relative to significant words. This means that, in Table 7, the WM = 29 for all n-grams but, with the tri-gram weight 1/(T i *Ln(T i )), is smaller than the 5-gram weight 1/T i, and 1/T i is also smaller than Ln(T i )/T i of 7-grams and 9-grams. This is not happening in the Vietnamese weighted frequency model in Table 8 because the value of WM is very large, 67. The weight 1/T i is applied to this Vietnamese model and to the corresponding 5-gram English model in Table 7. For example, we consider the following Vietnamese syllabic 9-gram, its closest English phrase - Vietnamese is a syllabic language - is only a word 5-gram. Therefore, because many English 5-grams will correspond to Vietnamese syllabic 7-grams or 9-grams, this Vietnamese model has the same weights as the 5-gram English weighted frequency model.

45 A Posteriori Individual Word Language Models for Vietnamese Language 117 Table 8. The weighted frequency model (Vietnamese QUB.) Sentence/QUB n-gram Perplexity Improvement (WM, Function) tri-gram % (67, 1/T i ) 5-gram % (67, 1/T i ) 7-gram % (67, 1/T i ) 9-gram % (67, 1/T i ) Table 9. The exponential decay model (Vietnamese QUB.) Sentence/QUB n-gram Perplexity Improvement (Decay, Cache) tri-gram % (150, 150) 5-gram % (150, 145) 7-gram % (150, 145) 9-gram % (150, 145) The best performance of the frequency models is shown by Table 10 and Table 11. Table 10. Improvement in perplexity for frequency models, all in sentence contexts (English WSJ.) Models tri-gram 9-gram Best Values Global 0.00% 26.77% Linear Interpolation 15.77% 34.32% λ=0.003, WM=29 Exponential Decay 8.13% 30.78% Decay=150, Cache=115 Weighted Frequency 22.87% 38.00% WM=29, Ln(T i )/T i Weighted Exponential Decay 22.84% 37.95% Decay=100, Cache=85, 1/T i Table 11. Improvement in perplexity for frequency models, all in sentence contexts (Vietnamese QUB.) Models tri-gram 9-gram Best Values Global 0.00% 20.88% Linear Interpolation 22.72% 36.20% λ=0.002, WM=46 Exponential Decay 18.19% 47.32% Decay=150, Cache=145 Weighted Frequency 30.15% 41.06% WM=67, 1/T i Weighted Exponential Decay 30.10% 41.05% Decay=100, Cache=100, 1/T i

46 118 Le Quan Ha et al. In Table 10 and Table 11, the best model in the Vietnamese corpus is the exponential decay model, which is different from the best in the English corpus because the Vietnamese language is formed by a more limited number of syllables than English words; hence, a Vietnamese syllable has many more different meanings than an English word. Only a significant Vietnamese syllable appearing in the immediate context of a target syllable can change the meaning of the target into a total different topic, but this depends much on the distance from the significant syllable model to the target. In the weighted frequency models, the best weight in English is (Ln(T i )/T i ) and Vietnamese (1/T i ). They are different because a Vietnamese sentence is generally longer in syllable length than its corresponding English sentence in word number, for example, the following sentence has 12 syllables and its corresponding English sentence that means I do research in individual word language models only contains 8 words. 7.3 Results for A Posteriori Models We investigated our new approach for calculating a posteriori probabilities using five models: the linear interpolation probability model, the exponential probability model, the weighted probability model, the linear interpolation frequency model, and the weighted frequency model. We found that the best performance was provided by the a posteriori weighted frequency model, which gave a 44.46% English improvement and 53.34% Vietnamese improvement in perplexity. This is better than the performance of much more computationally intensive methods based on clustering (Iyer & Ostendorf, 1999; Clarkson et al., 1997). The a posteriori weighted frequency model s results are displayed in Table 12 and Table 13, and the best results for all different a posteriori models are shown in Table 14 and Table 15. In order to compare to a priori models in Table 10, in the a priori weighted frequency model for English, the weights are different, namely Ln(T i )/T i in the a priori model and 1/T i *Ln(T i ) in the a posteriori model. Besides, the WM values, the number of significant word language models or m in Equation (14), are quite different, with values of 29 for the a priori and 16 for the a posteriori models. This can be explained as the concept that people can catch the meaning of a target word more clearly and quickly when they not only hear its previous words but also listen to its following words. Then, the English a posteriori model needs to increase weight on significant word models to the global model weight, but it needs to hear fewer significant words in the context before it can catch the meaning.

47 A Posteriori Individual Word Language Models for Vietnamese Language 119 Table 12. The a posteriori weighted frequency model (English WSJ.) Whole Sentence Perplexity Whole Sentence Improvement Weight tri-gram 5-gram 7-gram 9-gram tri-gram 5-gram 7-gram 9-gram T i *Ln(T i ) % 25.59% 27.53% 27.66% T i % 25.69% 27.62% 27.75% T i /Ln(T i ) % 25.80% 27.72% 27.85% Sqrt(T i ) % 27.13% 28.93% 29.03% Ln(T i ) % 30.30% 31.86% 31.93% Sqrt(Ln(T i )) % 30.65% 32.19% 32.26% Ln(1+Ln(T i )) % 30.77% 32.30% 32.36% 1/Ln(1+Ln(T i )) % 31.28% 32.78% 32.83% Sqrt(1/Ln(T i )) % 31.41% 32.89% 32.95% 1/Ln(T i ) % 31.81% 33.27% 33.32% Sqrt(1/T i ) % 38.24% 39.32% 39.33% Ln(T i )/T i % 43.32% 44.06% 44.02% 1/T i % 43.68% 44.37% 44.32% 1/T i *Ln(T i ) % 43.87% 44.52% 44.46% Table 13. The a posteriori weighted frequency model (Vietnamese QUB.) Whole Sentence Perplexity Whole Sentence Improvement Weight tri-gram 5-gram 7-gram 9-gram tri-gram 5-gram 7-gram 9-gram T i *Ln(T i ) % 19.62% 20.98% 21.22% T i % 19.74% 21.10% 21.33% T i /Ln(T i ) % 19.90% 21.25% 21.48% Sqrt(T i ) % 22.25% 23.51% 23.72% Ln(T i ) % 28.58% 29.64% 29.82% Sqrt(Ln(T i )) % 29.39% 30.42% 30.60% Ln(1+Ln(T i )) % 29.63% 30.65% 30.83% 1/Ln(1+Ln(T i )) % 30.92% 31.91% 32.08% Sqrt(1/Ln(T i )) % 31.19% 32.17% 32.34% 1/Ln(T i ) % 32.19% 33.14% 33.31% Sqrt(1/T i ) % 43.61% 44.32% 44.45% Ln(T i )/T i % 51.67% 52.22% 52.32% 1/T i % 52.55% 53.07% 53.16% 1/T i *Ln(T i ) % 52.74% 53.25% 53.34%

48 120 Le Quan Ha et al. Table 14. Improvements of a posteriori models (English WSJ.) A Posteriori Models tri-gram 9-gram Best Values Global 0.00% 26.77% Linear Interpolation Probability 30.99% 44.20% λ=0.2 Exponential Decay Probability 28.66% 42.82% Decay=100, Cache=5 Weighted Probability 30.82% 44.17% Sqrt(1/Ln(T i )) Linear Interpolation Frequency 25.17% 40.66% λ=0.001 Weighted Frequency 32.21% 44.46% WM=16, 1/T i *Ln(T i ) Similar to a posteriori Vietnamese models in Table 15, the weighted frequency model has WM=34 and a weight of 1/(T i *Ln(T i )), while the a priori ones are much larger, WM=67 and weights 1/T i. Table 15. Improvements of a posteriori models (Vietnamese QUB.) A Posteriori Models tri-gram 9-gram Best Values Global 0.00% 20.88% Linear Interpolation Probability 37.44% 47.63% λ=0.3 Exponential Decay Probability 40.31% 49.28% Decay=15, Cache=99 Weighted Probability 37.62% 47.34% Ln(T i ) Linear Interpolation Frequency 36.21% 47.03% λ=0.001 Weighted Frequency 45.70% 53.34% WM=34, 1/T i *Ln(T i ) Previously, for similar experiments, Sicilia-Garcia et al. (2005) reported improvements of 68%-69% for 7-gram models. Nevertheless, that was based on the flawed calculation previously described in the section, A Posteriori Method, so the results we present here can be viewed as a true reflection of the performance that can be achieved by these models. Nowadays, with the current condition of PCs, the programming issues of Sicilia-Garcia in training and accessing a language model of 9-grams no longer exist. With a 9-gram phrase length, on the global database of WSJ 4.5 gigabytes, we now complete the a posteriori probability linear interpolation model in 30 minutes, while Sicilia-Garcia finished this computer execution for 7-grams in 4 days nonstop. For the a posteriori weighted frequency model, we now execute only over 2 days for 9-grams, while Sicilia-Garcia completed 10 full days for 7-grams. (This is without speeding up the algorithms; it is only by upgrading to newer computers.) The different nature of the Vietnamese syllabic language causes a much smaller proportion of syllables to be significant than English significant words. In a Vietnamese sentence and an English sentence of the same length, there are considerably fewer significant syllables occurring, hence, inconsistent effects are found from our results.

49 A Posteriori Individual Word Language Models for Vietnamese Language Speech word-error-rate Results for A Posteriori Models We employ a Vietnamese large vocabulary continuous speech decoder linking to the five Vietnamese a posteriori models. Our speech recognition system is syllable-based HMM trained with Vietnamese speech data of 97,850 spoken statements extracted from the Vietnamese QUB text corpus; they are recordings of 326 speech hours by 60 Vietnamese speakers, including 30 men and 30 women from 18 years old to 51 years old. The number of states per HMM is 5, and each state was modeled using a mixture of 16 Gaussians. It is a tone-dependent phoneme model Vietnamese VnIPh Lexicon In March 2009, Ha created the Vietnamese lexicon called Vietnamese International Phonetic Lexicon version 1.0 or VnIPh lexicon ( We employ this lexicon in our speech recognition system; it includes an automatic lexicon generator to create 12,165 Vietnamese syllable entries. We would like to show a few Vietnamese syllables from this lexicon in Table 16. Table 16. The Vietnamese VnIPh Lexicon (a few entries) Vietnamese Syllable Phoneme Meaning in English In Table 16, phonemes 0, 1, 2, 3, 4, and 5 are the six tones of Vietnamese speech: level, low, high rising glottalized, dipping-rising, high rising, and low glottalized tones. In fact, there are 43 context-independent phonemes: 0, 1, 2, 3, 4, 5, AA, AH, AO, AW, AY, B, CH, D, EH, ER, EY, F, G, HH, IH, IY, K, L, M, N, NG, OW, OY, P, R, S, SH, SIL, T, UH, UW, V, W, WW, WY, Y, and Z Vietnamese Speech Tests We set up our own Vietnamese speech tests that included 21,451 Vietnamese syllables of 3,363 types in 3,834 statements. In Vietnamese speech, the number of syllables is countable because, using the Vietnamese spelling rule, all of the syllables can be obviously formed from the vowels, consonants, and tones. Hence, similar to English words, the number of Vietnamese words, which are combinations of syllables, is uncountable. Nevertheless, all of the possible Vietnamese syllables are fully stored in our lexicon and there are no

50 122 Le Quan Ha et al. out-of-vocabulary syllables in our Vietnamese speech tests, even though there are still unknown syllable combinations or unknown Vietnamese words out of the training corpus. Although the Vietnamese QUB training corpus is clean, without any external noises, our Vietnamese speech tests are noisy with all insertion/deletion/substitution cases of phonemes. Noisy spoken tests are recorded in realistic environments, around cars, fans, and other people talking; each utterance is recorded in a high, medium, or low noise condition. We apply an unconstrained phone recognizer, and the observed phone sequences are obtained. Our expected phone sequences are constructed using lexicon concatenation based on the Vietnamese syllable transcriptions. In comparison to the expected sequences, our observed sequences show three experimental conditions: well-matched, medium-matched, and highly-mismatched. In our tests, we have three common types of phone errors, which are 16.57% insertion, 18.24% deletion, and 16.21% substitution Vietnamese Language Models Our Vietnamese recognition baseline system uses a Katz back-off trigram language model; hence, we can calculate the word error rate (WER) as follows in Table 17. All of our Katz probabilities of Vietnamese phrases were stored in a hashed file inside the decoder. In order to link our individual word models into this decoder, we first calculated the combined probabilities of all n-grams from unigrams and bigrams up to 9-grams via the five models in Equations (6), (7), (8), (12), and (14). This took us ten days; then, we created five hashed files storing these Vietnamese probabilities with the same structure of the existing Katz hashed file. In the next step, we replaced the Katz file by each of our five files or our individual word language models. In this way, when speaking, the Vietnamese speech decoder is naturally able to execute on-line. Our Vietnamese word-error-rates are done by tests with direct microphone input from six held-out Vietnamese native speakers, four men and two women, ranging from 21 years old to 52 years old. Table 17. Noisy Speech A Posteriori WER (Vietnamese QUB.) A Posteriori Models WER Improvement over Baseline Baseline 45.65% Linear Interpolation Probability 33.67% 11.98% Exponential Decay Probability 35.78% 9.87% Weighted Probability 33.33% 12.32% Linear Interpolation Frequency 33.44% 12.21% Weighted Frequency 28.87% 16.78%

51 A Posteriori Individual Word Language Models for Vietnamese Language Conclusions We have described the concept of using individual word models to improve language model performance. Individual word language models permit an accurate capture of the domains in which significant words occur, thereby improving the model performance. The results indicate that individual word models offer a promising and simple means of introducing domain information into an n-gram language model. Humans probably hear the sounds of several words spoken before using a form of human language model to make a sensible sentence from the sounds, particularly when there are corruptions. Therefore, the idea of using the a posteriori method to define the domain might be more appropriate than the a priori method. We believe that the use of multiple word-domains, which need only large amounts of relatively inexpensive disk space, models the domain environment of any piece of written or spoken text more accurately than any other domain method. Our Vietnamese word-error-rate measurement to test this theory by linking to a speech decoder is also very good. For our future work, we will apply a posteriori language models in automatic speech recognition for English and other languages. Acknowledgement We would like to thank Dr. Sicilia-Garcia, Dr. P. Donnelly, Dr. D. Stewart, Dr. P. Hanna, and Prof. F. J. Smith, OBE MRIA. Our heartfelt acknowledgement is sent for the Reviewers of CLCLP. Finally, I would like to appreciate my wife, Mrs. Thuy Vo, and sister, Ms. Giang Vo, for their encouragement to complete this paper during winter in Calgary. Reference Blei, D. M., Ng, A. Y., & Jordan, M. I. (2003). Latent Dirichlet allocation. Journal of Machine Learning Research, 3, DOI: /jmlr Clarkson, P. R., & Robinson, A. J. (1997). Language Model Adaptation Using Mixtures and an Exponentially Decaying Cache. IEEE International Conference on Acoustics, Speech, and Signal Processing, 2, Munich, Germany. Donnelly, P. (1998). A Domain Based Approach to Natural Language Modelling. PhD Thesis. Queen s University Belfast, Northern Ireland. Donnelly, P. G., Smith, F. J., Sicilia-Garcia, E. I., & Ming, J. (1999). Language Modelling With Hierarchical Domains. The 6 th European Conference on Speech Communication and Technology, 4, Budapest, Hungary. Good, I. J. (1953). The Population Frequencies of Species and the Estimation of Population Parameters. Biometrika, 40,

52 124 Le Quan Ha et al. Ha, L. Q., Stewart, D. W., Ming, J. & Smith, F. J. (2008). Individual Word Probability Models. Web Journal of Formal, Computational & Cognitive Linguistics (FCCL), Issue 10, Russian Association of Artificial Intelligence. Hofmann, T. (1999). Probabilistic Latent Semantic Indexing. Proceedings of the Twenty-Second Annual International SIGIR Conference on Research and Development in Information Retrieval (SIGIR-99). Hofmann, T. (1999). The Cluster-Abstraction Model: Unsupervised Learning of Topic Hierarchies from Text Data. IJCAI, Hofmann, T., Puzicha, J., & Buhmann, J. M. (1998). Unsupervised Texture Segmentation in a Deterministic Annealing Framework. IEEE Trans. Pattern Anal. Mach. Intell. 20, Iyer, R. M., & Ostendorf, M. (1999). Modeling Long Distance Dependence in Language: Topic Mixture Versus Dynamic Cache Models. IEEE Transactions on Speech and Audio Processing, 17(1), Jelinek, F., Mercer, R. L., & Bahl, L. R. (1983). A Maximum Likelihood Approach to Continuous Speech Recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence, 5, Katz, S. M. (1987). Estimation of Probabilities from Sparse Data for the Language Model Component of a Speech Recogniser. IEEE Transactions on Acoustic Speech and Signal Processing, 35(3), Kuhn, R., & De Mori, R. (1990). A Cache-Based Natural Language Model for Speech Recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence, 12(6), Lau, R., Rosenfeld, R., & Roukos, S. (1993). Trigger-based Language models: A Maximum entropy approach. IEEE International Conference on Acoustics, Speech, and Signal Processing, 2, Minneapolis, MN. O Boyle, P., Owens, M., & Smith, F. J. (1994). Average n-gram Model of Natural Language. Computer Speech and Language, 8, Paul, D. B., & Baker, J. M. (1992). The Design for the Wall Street Journal-based CSR corpus. The Second International Conference on Spoken Language Processing, Seymore, K., Chen, S., & Rosenfeld, R. (1998). Nonlinear Interpolation of Topic Models for Language Model Adaptation. The 5 th International Conference on Spoken Language Processing, 6, Sydney, Australia. Sicilia-Garcia, E. I. (2002). A Study in Dynamic Language Modelling. PhD Thesis. Queen s University Belfast, Northern Ireland. Sicilia-Garcia, E. I., Ming, J., & Smith, F. J. (2001). Triggering Individual Word Domains in n-gram Language Models. The 7 th European Conference on Speech Communication and Technology, 1, Aalborg, Denmark.

53 A Posteriori Individual Word Language Models for Vietnamese Language 125 Sicilia-Garcia, E. I., Ming, J., & Smith, F. J. (2002). Individual Word Language Models and the Frequency Approach. The 7 th International Conference on Spoken Language Processing, Denver, Colorado. Sicilia-Garcia, E. I., Ming, J., & Smith, F. J. (2005). A posteriori multiple word-domain language model. The 9 th European Conference on Speech Communication and Technology (Interspeech-Eurospeech), Lisbon, Portugal.

54 126 Le Quan Ha et al.

55 Computational Linguistics and Chinese Language Processing Vol. 15, No. 2, June 2010, pp The Association for Computational Linguistics and Chinese Language Processing Improving the Template Generation for Chinese Character Error Detection with Confusion Sets Yong-Zhi Chen, Shih-Hung Wu, Ping-che Yang +, and Tsun Ku + Abstract In this paper, we propose a system that automatically generates templates for detecting Chinese character errors. We first collect the confusion sets for each high-frequency Chinese character. Error types include pronunciation-related errors and radical-related errors. With the help of the confusion sets, our system generates possible error patterns in context, which will be used as detection templates. Combined with a word segmentation module, our system generates more accurate templates. The experimental results show the precision of performance approaches 95%. Such a system should not only help teachers grade and check student essays, but also effectively help students learn how to write. Keywords: Template Generation, Template Mining, Chinese Character Error. 1. Introduction In essays written in Chinese by students, incorrect Chinese characters are quite common. Since incorrect characters are a negative factor in essay scoring, students should avoid such errors in their essays. Our research goal is to build a computer tool that can detect incorrect Chinese characters in student essays and correct them, so that teachers and students can learn faster with help from the computer system. Compared with the detection of spelling errors in English, the detection of incorrect Chinese characters is much more difficult. In English, a word consists of a series of letters while a meaningful Chinese word usually consists of 2 to 4 Chinese characters. The difficulty lies partly in the fact that there are more than 5,000 high-frequency characters. In previous works on Chinese character error detection systems (Zhang, Huang, Zhou, & Department of Computer Science and Information Engineering, Chaoyang University of Technology { , shwu}@cyut.edu.tw The author for correspondence is Shih-Hung Wu. + Institute for Information Industry {maciaclark, cujing}@iii.org.tw

56 128 Yong-Zhi Chen et al. Pan, 2000) (Ren, Shi, & Zhou, 1994), a confusion set for each character is built and is used to detect the character error with the help of a language model. The confusion set is based on a Chinese input method. The characters that have similar input sequences probably belong to the same confusion set. For example, the Wubizixing input method (Wubi), which is a Chinese character input method primarily for inputting both simplified and traditional Chinese text in a computer, is used in (Zhang, Huang, Zhou, & Pan, 2000). The Wubi method is based on the structure of the characters rather than on the pronunciation. It encodes every character in four keystrokes at the most. Therefore, if one keystroke is changed, another character similar to the correct one will show up. Once a student chooses the similar character instead of the accurate one, a character error is established, and a confusion set is automatically generated by the character error. Another approach is to manually edit the confusion set. Common Errors in Chinese Writings gives 1477 common errors (National Languages Committee, 1996). Nevertheless, this amount is not sufficient to build a system. Hung manually compiled 6701 common errors from different sources (Hung & Wu, 2008). These common errors were compiled from essays of junior high school students and were used in Chinese character error detection and correction. Since the cost of manual compilation is high, Chen et al. proposed an automatic method that can collect these common errors from a corpus (Chen, Wu, Lu, & Ku, 2009). The idea is similar to template generation, which builds a question-answer system (Ravichandran & Hovy, 2001) (Sung, Lee, Yen, & Hsu, 2008). The template generation method investigates a large corpus and mines possible question-answer pairs. Templates for Chinese character error detection can be generated and tested by the chi-square test on the basis of a large corpus. In this paper, we will further improve the methods for building confusion sets and automatically generating a template. According to recent studies(liu, Tien, Lai, Chuang, & Wu, 2009a; 2009b), character errors in student essays are of four major types: errors in which characters have similar shapes (30.7%), errors in which characters have similar pronunciation (79.9%), errors in which the two previous types are combined (20.9%), and other errors (2.4%). Therefore, an ideal system should be able to deal with these errors, especially those resulting from similar pronunciation and similar character shapes. The confusion set for similar pronunciation is relatively easy to build, whereas the confusion set for similar shapes is more difficult. In addition to the Wubi input method, the Cangjie input method is also used to compile confusion sets (Liu & Lin, 2008). The paper is organized as follows. In Section 2, we introduce the system design and related works. In Section 3, we describe a new process of template generation. Section 4 describes the experimental procedure and the data. Finally, in Section 5, we give the conclusion and propose our future research.

57 Improving the Template Generation for 129 Chinese Character Error Detection with Confusion Sets 2. System Design 2.1 Chinese Character Error Detection and Correction System The system that can detect and correct Chinese character errors works as follows. First, it needs a student to input an essay. The system then reports the errors in the essay and gives suggestions on correction, as shown in Figure 1. Such a system uses templates that can detect whether common errors have occurred. A template consists of a pair of words, a correct one and an error one, such as 辯論會 - 辨論會. For example, if the error template 辨論會 is matched in an essay, our system can conclude that there is an error and make a suggestion on correction to 辯論會. Figure 1. System function of Chinese character error detection in an essay In previous works, these templates were compiled manually (Liu, Tien, Lai, Chuang, & Wu, 2009b). The quality of the manually-edited templates is high. Nevertheless, the method is time-consuming and costs too much manpower. Therefore, an automatic template generation method based on the context of errors was proposed in 2009 (Chen, Wu, Lu, & Ku, 2009), several examples of automatically generated tri-gram and four-gram templates are shown in Figure 2. The automatic template generation method is less costly; however, it does not accommodate conventional vocabulary. The template generation method has a serious drawback. In Figure 2, we find that several templates contain unrecognizable words, such as 辯護律, 視辯論, and 電視辯, which are trigrams of Chinese characters that do not have

58 130 Yong-Zhi Chen et al. any meaning. These templates can be used to detect character errors, but are not suitable for suggesting corrections. In the following subsections, we will propose a new method to avoid this drawback. Templates Templates Correct Error Correct Error 會首長 會首常 清潔隊長 清潔隊常 會給予 會給于 交通隊長 交通隊常 辯論會 辨論會 辯護律師 辨護律師 辯護律 辨護律 視辯論會 視辨論會 的辯論 的辨論 政策辯論 政策辨論 視辯論 視辨論 電視辯論 電視辨論 電視辯 電視辨 公開辯論 公開辨論 半世紀 辦世紀 半個世紀 辦個世紀 半以上 辦以上 一年半的 一年辦的 半個小 辦個小 的另一半 的另一辦 Figure 2. The templates for error detection and correction in (Chen, Wu, Lu, & Ku, 2009) 2.2 Confusion Set The first step in template generation is to replace one character in a word with a character in the corresponding confusion set. For example, by replacing one character in the correct word 芭蕉, we get a wrong word 笆蕉. Such a correct-wrong word pair is used as the template for error detection and correction suggestion. According to Liu et al. (Liu, Tien, Lai, Chuang, & Wu, 2009a; 2009b), the most common error types are characters with similar shapes and characters with similar pronunciation. The percentage of these two types of errors combined is 89.7% of all errors. Therefore, the confusion set should deal with characters with similar pronunciation and shapes. We first compile all of the characters that have the same pronunciation from a dictionary and make them the elements of a confusion set. For example, 八 (ba1) and 巴 (ba1) have the same pronunciation. Therefore, they belong to the same confusion set. To reduce the size of the confusion set, we treat characters with different tones as belonging to different sets, even though they sound similar. For example, 罷 (ba4) is not in the confusion set of 八 (ba1). We formed 1,351 sets with a total of 15,160 characters, as shown in Figure 3. In this paper, we use a simple rule to compile characters with similar shapes. In the first book on Chinese characters, known as Shuowen Jiezi ( 說文解字 ) (Xu, 2009), in the second

59 Improving the Template Generation for 131 Chinese Character Error Detection with Confusion Sets century, radicals ( 部首 ) were used to categorize characters. We use the key component of a character, its radical, as the basic shape of the character to find the characters with the same radicals. There are 214 radicals in Chinese, according to the Kangxi Dictionary ( 康熙字典 ) (Zhang, 1999). Therefore, we compile 214 confusion sets with a total of 9,752 different characters. Figure 4 shows some examples. After constructing the confusion sets, our system can find characters with the same pronunciation and characters with similar shapes for any character that is input. For example, given a character 兇, the system can find characters with the same pronunciation 凶兄匈洶 恟胸, and characters with similar shapes 兄光兆先兌克免, as shown in Figure 5. This is a crucial step of our new template generation. Zhuyin Pinyin Characters ㄅㄚ ba1 蚆扒八巴仈叭朳芭疤捌笆粑豝鈀吧 ㄅㄚˊ ba2 鈸茇拔胈跋菝詙軷魃鼥犮 ㄅㄚˇ ba3 鈀把靶 ㄅㄚˋ ba4 伯罷霸猈弝爸壩灞把耙 ㄅㄛ bo1 剝嚗波袚玻柭砵缽啵菠碆撥嶓蹳鱍岥播襏 ㄅㄛˊ bo2 爆伯犮襏挀蔔柏瓝薄泊謈濼鋍帛勃胉挬浡 ㄅㄛˇ bo3 簸跛蚾 ㄅㄛˋ bo4 播檗蘗亳擘譒北挀薜簸繴 Figure 3. Examples of characters in confusion sets Radicals 一丶丿乙亅二亠人儿 Characters 一丂丁七三下丈上万丌丑丐不丏丙世丕且丸凡丹主乂乃久么之尹乍乏乎乒乓乑乖乘乙九乜也乞乣乩乳乾亂了予事二于云井互五亓亙些亞亟亡亢交亦亥亨享京亭亮亳亶亹人仁什仃仆仇仍今介仄仂仉以付仔仕他仗兀元允充兄光兇兆先兌克免兕兔兒兗党兜 入 入內全兩 Figure 4. Examples of characters in confusion sets

60 132 Yong-Zhi Chen et al. Figure 5. Combination of the two confusion sets for a given character 2.3 Automatic Template Generation Figure 6 shows the flowchart of our automatic template generation process. The basic assumption is that the corpus might contain more correct words than wrong ones. Therefore, our system first replaces one character in the correct words to form the corresponding wrong words. Then, our system checks the frequency of the words in the corpus. If the replacement creates a word with a relatively high frequency, we do not treat it as a wrong word. Figure 6. The flowchart of the automatic template generation process As we mentioned in Section 2.1, the automatically-generated templates might not be suitable for suggesting corrections. To overcome this drawback, we use existing vocabulary, instead of n-gram character sequences, as the candidate for a template. There are 145,608 words in the MOE dictionary (Ministry of Education, 2007). We treat them as the seeds of the templates. In our experiment, we focus on 4,998 high-frequency characters that were compiled on the basis of a 1998 survey (National Languages Committee, 1998). Our system generates templates by checking each high-frequency character and finding all of the words that contain the character. Then, the system replaces the character in each

61 Improving the Template Generation for 133 Chinese Character Error Detection with Confusion Sets word with a character in the corresponding confusion set. The correct-wrong word pair undergoes a simple statistical test. If it passes the test, it will be kept as a template; otherwise, it will be discarded. The statistical test is based on the frequency of each word in the pairs appearing in a large corpus. To prevent the process from generating controversial templates, our system also conducts a close test. The close test checks whether the new template will cause a false alarm on our old test data. The template that generates conflicting templates will also be discarded. The close test threshold is set to 0, which means any template that might cause a false alarm will not be used. A template generation example is shown in Figure 7. Figure 7. A template generation example, where two templates are generated for an input character 官. The statistical test in our system is not a rigid test. We tune the threshold of relatively high frequency based on two formulae. One is adopted from the chi-square test, and the other one is from our observation. The first test is a simplified (n=1) chi-square test used in a previous work (Hung & Wu, 2008): 2 O E =, (1) 2 ( ) X E where E is the frequency of a correct word and O is the frequency of a wrong word. To avoid further disputation, we assume that E>O in our study. The chi-square test provides a threshold mechanism to decide whether a correct-wrong pair is a proper template or not. In this study, we suggest the test should be like Equations (2) and (3). Cfreq > Wfreq, Cfreq > AverageFreq (2) n Cvocabulary() i i= 1 Threshold = n, (3)

62 134 Yong-Zhi Chen et al. where Cfeq is the frequency of the correct word, Wfeq is the frequency of the wrong word, and AverageFreq is the average of the frequencies of all correct words. If the frequency of the correct word is higher than the threshold and if the square root of the frequency of the correct word is higher than the frequency of the wrong word, then the pair passes the test. We have found that the templates that do not pass the test are also the ones that will cause false alarms; for example, the pairs 未來 - 為來, 已經 - 以經, and 但是 - 但事. When the context is different, these templates do not always give correct detection results and cause false alarms. 2.4 Word Segmentation As in the examples above, short templates with only two characters could cause false alarms. The reason is that, when we treat words as bi-gram character sequences, many word boundaries may be unclear. For example, as shown in Figure 8, the template 擁有 - 雍有 can be used to detect and correct the first sentence, 一個人可雍有很多快樂, in which one of the word pair appears, but the template 擁有 - 以有 cause a false alarm in the second sentence, 一個人可以有很多快樂. We find that this failure can be avoided by using correct word segmentation. The character 以 should be a part of the previous word 可以. If we have enough confidence in the word segmentation, then the characters in a segmented word should not be candidates for character error detection. Figure 8. A false alarm in the second sentence for a short template 擁有 - 雍有 and 擁有 - 以有 We assume that a word segmentation tool can give the correct results for normal input sentences and does not segment sentences with wrong character sequences into words. Figure 9 shows the segmentation results of the two sentences shown in Figure 8. In our experiment, we used the segmentation tool provided by CKIP, Academia Sinica 1. With the help of this segmentation tool, our system can compile more accurate short templates. Some short templates are shown in Figure

63 Improving the Template Generation for 135 Chinese Character Error Detection with Confusion Sets Figure 9. Segmentation tool can help prevent false alarms Correct templates Incorrect templates Correct templates Incorrect templates Correct templates Incorrect templates 衝擊 衝急 絆腳石 伴腳石 逼不得已 逼不得己 檢視 機視 大部分 大不分 情非得已 情非得巳 經濟 經紀 手電筒 手電桶 逼不得已 逼不得巳 循環 循還 不經意 不經易 大勢已去 大勢以去 成績 成積 不願意 不願易 不能自己 不能自以 薪水 新水 董事長 懂事長 迫不得已 迫不得以 賺錢 購錢 三輪車 三軸車 情非得已 情非得以 關鍵 關建 腦震盪 腦振盪 萬不得已 萬不得以 老闆 老版 辦公室 辨公室 逼不得已 逼不得以 雖然 隨然 成績單 成積單 巡弋飛彈 巡曳飛彈 Figure 10. Some short templates generated by our system 3. Experimental Settings, Results and Analysis 3.1 Training Corpus and Student Essays Our method requires a large corpus to compile templates. Therefore, we used the largest available news corpus as our training set. The corpus is described in Table 1. Table 1. Corpus statistics Year News sources # of Docs File size China Times 38,163 China Times Commercial 25, China Times Express 5, MB Central Daily News 27,770 China Daily News 34,728

64 136 Yong-Zhi Chen et al United Daily News 249, MB United Daily News 172, United Express 91,958 Min Sheng News 168,807 Economic Daily News 463, GB Student essays were collected from one junior high school in Taipei. We used some of the essays for the close test and the rest as the open test, keeping them unseen to the system. The students were 7 th or 8 th graders. The essays were reviewed by their teachers, and the character errors were highlighted. These 3264 essays were written by hand and were digitized later. See Figure 11 for an example. This is part of our experimental setting that tries to avoid the influence of different input methods. We deleted some symbols and characters that could not be represented by Unicode. Figure 11. The file format of our test corpus Table 2 shows the analysis of the student essays. Most of the characters (94%) in use fell into the frequent characters set. Character errors were not very serious for most of the students, with less than 2 character errors per essay. Table 3 shows our analysis of the character error types. We find that even in written essays, students tend to write characters having the same pronunciation (66~70%). There is also a high percentage of wrong written characters with the same radical (13~16%). Table 4

65 Improving the Template Generation for 137 Chinese Character Error Detection with Confusion Sets shows the templates most used for the student essays. These templates are quite common and are too simple for teachers to teach at the 7 th and 8 th grade levels. A system that can correct these errors may reduce the work of teachers. Table 2. Analysis of the student essays # of Essays Average score Average # of characters Average # of character errors % of frequent characters Close test essay % Open test essay % Table 3. Analysis of the character error types in student essays % with the same radical % with the same pronunciation % of both % out of the two main types Close test essay 13.82% 70.27% 4.92% 20.81% Open test essay 16.96% 66.31% 2.85% 19.58% Table 4. The most used templates in the test corpus Close essay Open essay Correct 已經變得自己景象一旦寄託已經畢竟而已根本 Wrong 己經變的自已景像一但寄托以經必竟而己跟本 Correct 自己一旦已經選擇煩惱應該已經而已選擇後悔 Wrong 自已一但己經選則煩腦因該以經而己撰擇後侮 3.2 System Evaluation In this study, we compare the quality of characters manually compiled from books and students with that of automatically generated ones. Since the frequencies of 2-character words, 3-character words, and 4-character words are very different, our system uses different thresholds , 500, and 100 for 2-character words, 3-character words, and 4-character words, respectively, in the experiment. The precision and recall are defined as follows: dr dr ( ) ( ) Macro Recall = r (4) Macro Precision = sd (5) N N (dr) (dr) Micro Recall = (6) Micro Precision = (7) (r) (sd) where dr is the number of correct characters, r is the number of character errors, sd is the number of character errors that our system detects, and N is the number of all of the essays.

66 138 Yong-Zhi Chen et al. Macro Precision and Macro Recall are focused on the performance of correction per essay. This is what real world students might encounter with the system. As Micro Recall and Micro Precision treat the whole data set as one essay, they are suitable for evaluating the average performance of the system. We prefer high precision while maintaining a relatively high recall because we do not want the users to see too many false alarms. 3.3 Experimental Results We conducted a series of experiments to determine how to improve our system. First, we used confusion sets and the chi-square test to generate templates and compared the performance with the previous work, which did not use confusion sets. Second, we tested whether the square root test is more suitable for our system than the chi-square test. Third, we tested the influence of the segmentation added to our system. We report the best performance of the experimental results by combining the automatically generated templates with the manually edited templates The Comparison of Eexperimental Results of Four Automatic Template Generation Settings Figure 12 shows the experimental results of using the chi-square test in template generation. Setting A used the automatically generated 19,402 templates in the previous work. Setting B used the confusion sets during the process of automatic template generation. The total number of generated templates was 54,253. The performance of the method proposed in this paper is better than the previous work for both precision and recall. Setting C was the automatically generated templates using the confusion set and the square root test. The total number of templates was 50,467. This new setting results in much higher precision. The Macro Precision value is even better than the manually edited Macro Precision value. This result shows that, when we reduce the automatically generated templates with the square root test, we also reduce noise. For Setting D, our system used confusion sets and a word segmentation tool before the square root test, which generated 9,013 templates. We find that the number of templates is reduced while the performance is improved in terms of both Macro Precision and Micro Precision. The trade off is the performance of recall.

67 Improving the Template Generation for 139 Chinese Character Error Detection with Confusion Sets 100% 80% 60% 40% 20% 0% A:Previous work B:Chi-squre test C:square root test D:square root+word segmentation 95.55% 91.27% 84.31% 80.10% 73.58% 62.60% 49.55% 32.69% 32.07% 33.95% 28.10% 29.24% 10.52% 10.08% 6.05% 7.43% Macro Precision Macro Recall Micro Precision Micro Recall Figure 12. The comparison of experimental results of four automatic template generation settings Combining Automatically Generated Templates with Manually Edited Templates Figure 13 shows the comparison of the performance of our system combing automatically generated templates with manually edited templates. Setting E used the 6,701 manually edited templates. Setting F used the combination of Setting E and Setting C, which had a total of 57,167 templates. Setting G used the combination of Setting E and Setting D, totaling 15,713 templates. The performance of the combinations declines a little bit in terms of both Macro Precision and Micro Precision. Nevertheless, there is an increase in both Macro Recall and Micro Recall. Compared with the results in the previous experiment, the combination helps the overall performance. This means that our system can incorporate more templates and attain better performance in the future.

68 140 Yong-Zhi Chen et al. E:Manually Edited F:Manually Edited+C G:Manually Edited+D 100% 80% 60% 40% 20% 89.21% 87.63% 84.14% 43.31% 38.32% 40.01% 69.94% 68.73% 65.86% 23.63% 17.62% 19.48% 0% Macro Precision Macro Recall Micro Precision Micro Recall Figure 13. A comparison of experimental results of combining manually edited with automatically generated templates Based on the analysis of the confusion sets, our system should have a 70% to 80% recall rate because we compile all of the characters with the same pronunciation and some similar characters in the confusion sets. Nevertheless, the recall remains low, even though we are able to control the high-precision performance. Therefore, we will need to conduct further analysis of our system. 3.4 Analysis of the Mistakes in the Experiment In this subsection, we discuss the 90,135 templates in Setting I of the third experiment, which were generated by using confusion sets, word segmentation, and the square root test. This setting was designed to maintain high precision and to increase recall Regarding the Precision Theoretically, our system can get 100% precision using templates. In practice, however, there are still many exceptions. In Table 5, we list some false alarms in the open tests. According to an online dictionary (Ministry of Education, 2007), some templates that we compiled are interchangeable, such as: 垃圾桶 - 垃圾筒, 奇蹟 - 奇跡, 電線桿 - 電線杆, and 銷聲匿跡 - 消聲匿跡. This is not consistent with the judgment of some teachers. Some templates are just too short and cannot include the necessary context in order for a correct decision to be made, such as 一再 - 一在. The necessary context should include more semantic rather than surface syntax. There were some bad templates that our system should

69 Improving the Template Generation for 141 Chinese Character Error Detection with Confusion Sets have not generated, such as 放聲大哭 - 放聲大叫, 不用說 - 不用講, and 讀書人 - 讀書做, which can be attributed to the size of the corpus. Nevertheless, no corpus is large enough to be perfect for all applications. We find that these are the major causes of false alarms. Table 5. Some templates that caused false alarms Correct word 垃圾桶奇蹟電線桿銷聲匿跡一再放聲大哭不用說讀書人 Wrong word 垃圾筒奇跡電線杆消聲匿跡一在放聲大叫不用講讀書做 Regarding the Recall We treated the errors that the teachers provided from the student essays as templates and compared them to the automatically generated templates, as shown in Table 6. The first column shows the percentage of not in the automatically generated template. The second column shows the percentage of an error occurring in a word that is not in the dictionary. The third column shows the percentage of an error occurring in a word that is not in the corpus. The last column shows the percentage of an error occurring in a word that is neither in the dictionary nor in the corpus. We find that most student errors were not mined from the news corpus, although our system has mined many useful error templates. From the union set of those not in a dictionary and not in a corpus, we find that 53.17% of the necessary templates in the close test set cannot be generated by our system, while 32.97% of the necessary templates in the open test cannot be generated by our system. This is a mismatch of the corpus and student essays. The assumption of our system is that the corpus contains the correct and wrong usages. Nevertheless, since news reporters and junior high school students make character errors for different words, we need to have a more suitable corpus to improve our system. If we have a more contemporary dictionary that includes the words in Table 7, our system can perform better. Table 6. Comparison of real world errors to system generated templates Not matching template Not in dictionary Not in corpus Neither in dictionary nor in corpus Close test essay 91.53% 37.73% 35.64% 20.20% Open test essay 93.15% 16.27% 23.94% 7.24% Table 7. New words not in dictionary 佈告欄 蒸飯機 值日生作業本辦派對睡午覺全班齊心勤加練習羞恥心無厘頭 重拾信心莽莽撞撞 淘汱 漆彈場偶像劇積陰德融入團體 芬多精 燒炭 拉筋

70 142 Yong-Zhi Chen et al. 4. Conclusion and future works Based on the confusion sets of Chinese characters, word segmentation, and the square root test, our system can generate a large number of templates from a corpus. These templates can detect and correct Chinese character errors in essays. The templates are more readable and have better performance in both precision and recall performance compared to that of previous system. To improve the system, we will work in two areas. In the knowledge part, we will enlarge the confusion sets to include more seeds for template generation. We will compile a more suitable corpus for detection and correction of errors in student essays. For the dictionary, we will collect more contemporary terms via the Internet, such as from Wikipedia and Wikitionary. For the language model part, we will use the student essays that we collected in this study to generate an error model, and use that error model to help determine character errors. Acknowledgement This study was conducted under the "Project Digital Convergence Service Open Platform" of the Institute for Information Industry, which is subsidized by the Ministry of Economic Affairs of the Republic of China. Reference Chen, Y.-Z., Wu, S.-H., Lu, C.-C., & Ku, T. (2009). Automatic Template Generation for Chinese Essay Spelling Error Detecting System. The 13th Global Chinese Conference on Computer in Education, Hung, T.-H., & Wu, S.-H. (2008). Chinese Essay Error Detection and Suggestion System. Taiwan E-Learning Forum. Liu, C.-L., & Lin, J.-H. (2008). Using structural information for identifying similar Chinese characters. Proceedings of the Forty Sixth Annual Meeting of the Association for Computational Linguistics: Human Language Technologies (ACL'08), Liu, C.-L., Tien, K.-W., Lai, M.-H., Chuang, Y.-H., & Wu, S.-H. (2009). Capturing errors in written Chinese words. Proceedings of the Seventh Workshop on Asian Language Resources (ALR7), the Forty Seventh Annual Meeting of the Association for Computational Linguistics (ACL'09), Liu, C.-L., Tien, K.-W., Lai, M.-H., Chuang, Y.-H., & Wu, S.-H. (2009). Phonological and logographic influences on errors in written Chinese words. Proceedings of the Seventh Workshop on Asian Language Resources (ALR7), the Forty Seventh Annual Meeting of the Association for Computational Linguistics (ACL'09), Ministry of Education. (2007). MOE Chinese Dictionary( 教育部重編國語辭典修訂本 ). Taiwan: Ministry of Education.

71 Improving the Template Generation for 143 Chinese Character Error Detection with Confusion Sets National Languages Committee. (1996). Common Errors in Chinese Writings ( 常用國字辨似 ). Taiwan: Ministry of Education. National Languages Committee. (1998). The Investigation on Common Used Word( 八十七年常用語詞調查報告書 ). Taiwan: Ministry of Education. Ravichandran, D., & Hovy, E. (2001). Learning surface text patterns for a Question Answering system. in Proceedings of the 40th Annual Meeting on Association for Computational Linguistics, Ren, F., Shi, H., & Zhou, Q. (1994). A hybrid approach to automatic Chinese text checking and error correction. In Proceedings of the ARPA Work shop on Human Language Technology, Sung, C.-L., Lee, C.-W., Yen, H.-C., & Hsu, W.-L. (2008). An Alignment-based Surface Pattern for a Question Answering System. the IEEE International Conference on Information Reuse and Integration, Xu, S. (2009). Shuowen Jiezi( 說文解字 ). Volumes publishing company( 萬卷出版公司 ). Zhang, Y.-s. (1999). Kangxi Dictionary( 康熙字典 ). Chung Hwa Book company( 中華書局 ). Zhang, L., Huang, C., Zhou, M., & Pan, H. (2000). Automatic detecting/correcting errors in Chinese text by an approximate word-matching algorithm. Proceedings of the 38th Annual Meeting on Association for Computational Linguistics,

72 144 Yong-Zhi Chen et al.

73 Computational Linguistics and Chinese Language Processing Vol. 15, No. 2, June 2010, pp The Association for Computational Linguistics and Chinese Language Processing 以最佳化及機率分佈標記形聲字聲符之研究 Annotating Phonetic Component of Chinese Characters Using Constrained Optimization and Pronunciation Distribution 張嘉惠 林書彥 李淑瑩 蔡孟峰 李淑萍 + 廖湘美 + 孫致文 + 黃鍔 # Chia-Hui Chang, Shu-Yen Lin, Shu-Ying Li, Meng-Feng Tsai, Shu-Ping Li, Hsiang-Mei Liao, Chih-Wen Sun, and Norden E. Huang 摘要一般說來, 漢字乃圖形文字, 無法像英文等拼音文字一樣, 一旦學會拼音方法, 即有基本的閱讀能力 相對的, 漢字讀寫的學習進展則相當緩慢, 而且必須搭配注音符號或是其他拼音方法, 才可知道每個漢字的發音 事實上漢字中有八成的字是形聲字, 形聲字不僅可由形旁表意, 又可以聲符表音, 因此即使沒見過的字也可以由偏旁推論其音及義 不過主要的困難在於聲旁未必一定同音, 可能是相近的發音, 之間的演變規則尚未有人探究過, 例如 : 泡 抱 飽三個字同樣與 包 的發音相近, 然而發音如何由 包 的發音轉變成其他三個字的發音, 則仍待研究 本論文首先嘗試以自動化方式判定漢字聲符, 做為研究形聲字發聲規則的第一步 實驗顯示, 我們所提的兩種方式, 發音相似度比較法在 9593 個形聲字中的判定聲符準確率為 90.7%, 而構件發聲分佈比較法則 國立中央大學資訊工程所, 台灣桃園縣中壢市中大路 300 號 Dept. of Computer Science and Information Engineering, National Central University, Taiwan chia@csie.ncu.edu.tw The author for correspondence is Chia-Hui Chang. + 國立中央大學中文系 Dept. of Chinese Literature, National Central University # 國立中央大學數據中心 Research Center for Adaptive Data Analysis, National Central University

74 146 張嘉惠等 可達到 98.1% 的準確率, 可以加速形聲字聲符標記所需的大量人力工作與時間 關鍵字 : 形聲字 聲符 發音相似度 最佳化 機率分佈 KL divergence 1. 簡介 Abstract Generally speaking, Chinese characters are graphic characters that do not allow immediate pronunciation unless they are accompanied with Mandarin phonetic symbols (zhuyin) or other pinyin methods (e.g. romanization system). In fact, about 80 to 90 percents of Chinese characters are pictophonetic characters which are composed of a phonetic component and a semantic component. Therefore, even if one had not seen the character before, one can make a logical guess at the character's pronunciation and meaning from its phonetic and semantic symbols. In order to analyze such relations, we start by analyzing the characteristics of phonetic components. We found two interesting features that could automatically identify the phonectic components of Chinese characters. One is pronunciation similarity, the other is pronunciation distribution. Experiments show that these two methods have high accuracy (90.8% and 98.1% for 9593 pictophonetic characters) in predicting the phonetic components of pictophonetic characters. These methods can save a lot of time and effort during the annotation of phonetic symbols in the early stage. Keywords: Picto-phonetic Compounds, Phonetic Component, Pronunciation Similarity, Pronunciation Distribution, Optimization. 漢語字形及音讀的繁複, 向為初學者及外籍人士所苦, 即使會說華語的海外華人對於漢字的認識也可能相當有限 最主要的原因在於漢字乃圖形文字 (pictograph system), 無法像英文等拼音文字 (alphabet system) 一樣, 一旦學會拼音方法 (phonetic representation), 即有基本的閱讀能力 相對的, 漢字讀寫的學習進展則相當緩慢, 而且必須搭配注音符號 (Chinese phonetic symbols) 或是其他拼音方法, 才可知道每個漢字的發音 這樣的限制, 對於漢字的學習相當不利, 這也是為什麼二十世紀初期許多中國革命家意欲將漢字拉丁化的主要原因 漢字的構成包含象形 指事 會意 形聲 轉注 假借 ( 總稱六書 ( 許慎,1988)), 其中象形 指事是 造字法, 會意 形聲是 組字法, 轉注 假借是 用字法 事實上形聲字所占的比例相當高, 約佔八 九成 形聲字不僅可由形旁表意, 又可以聲符表音, 因此即使沒見過的字也可以由偏旁推論其音及義, 所謂 有邊讀邊沒邊讀中間 即是此意 不過主要的困難在於聲旁僅代表相近的發音, 之間的演變規則尚未有人探究

75 以最佳化及機率分佈標記形聲字聲符之研究 147 過, 例如 : 泡 抱 飽三個字同樣與 包 的發音相近, 然而發音如何由 包 的發音轉變成其他三個字的發音, 則仍待研究 為了解漢字中形聲字與其聲符之間發音規則的轉變, 我們必須先知道每一個形聲字的聲符 由於形聲字與聲符發音相近, 因此我們憶測可用形聲字與其構件之間的發音相似度做為聲符預測方法, 因此我們一方面尋求聲母 韻母之間發音相似度, 一方面也建立一個形聲字源標記系統, 由中央大學中文所四位研究生與三位教授參與人工標記漢字構形資料庫中 有注音標示的漢字是否為形聲字以及其聲符構件, 做為預測方法的驗證 另外為了提升經由發音相似度比較法判斷聲符之準確率, 我們也採用限制性最佳化技術, 自動求得新的發音相似度分數 不過發音相似度比較法僅參考單一漢字的資訊, 事實上做為聲符構件的漢字, 其衍生字的發聲分佈比非聲符構件的漢字發聲分佈更為集中 因此我們進一步提出構件發聲分佈比較法, 經由計算每個構件的發聲分佈與所有漢字的發聲分佈的 KL 值, 做為此構件做為聲符的強度代表 實驗結果顯示, 發音相似度比較法在 9593 個已標示的形聲字中判定聲符準確率為 90.8%, 而構件發聲分佈比較法則可達到 98.1% 的準確率, 顯示兩種方法做為聲符判斷問題的可行性 雖然形聲字的聲符預測並非計畫最終目的, 而且在聲符標注完成後, 預測聲符的需求即消失不在, 但是透過發音相似度最佳化方法所得的聲母, 韻母相似度參數或許有助於未來漢字字音處理的研究, 同時部件發音強度也可做為漢字學習順序之參考, 仍有相當的重要性 2. 相關研究漢字的使用從殷商的甲骨文算起已達 3,400 年之久, 由於結構複雜, 因此文字學在中國特別發達 文字學的研究包括文字起源 發展 性質 體系, 文字的形 音 義的關係, 正字法以及個別文字演變的情況等等議題 為有系統的研究, 中央研究院資訊科學研究所文獻處理實驗室從 1993 年開始, 陸續建構古今文字的源流演變 字形結構及異體字表, 做為記錄漢字形體知識的資料庫, 也就是漢字構形資料庫 ( 莊德明 謝清俊,2005) 漢字構形資料庫不僅銜接古今文字以反映字形源流演進, 也記錄了不同歷史時期的文字結構 另外也由於開發漢字部件檢字系統, 得以解決缺字問題 然而過去的研究著重在字形知識的整理, 尚未涉及字音與字義的處理 ; 因此近年來開始文字學入口網站建置計畫 ( 莊德明 謝清俊,2005; 莊德明 鄧賢瑛,2008) 一如其文所述 : 漢字構形資料庫目前只著重在字形知識的整理, 尚未涉及字音與字義 ; 建立一個形 音 義俱備的漢字知識庫, 仍是我們長遠的目標 ", 因此本計畫 漢語系統音源脈絡之分析 " 的目的即是以挑戰漢字的發音規則知識庫為出發, 除了了解漢字發音規則外, 也希望藉由此項研究找出一套形聲字發音轉換規則, 讓華語學習可以在聲符與規則的輔助下, 順利讀出字的發音 為逹成此目的, 第一步我們必須了解每個漢字是否為形聲字,

76 148 張嘉惠等 以及了解形聲字聲符的部件, 進而解析聲符與最終發音之間轉換的規則 因此我們首先設計一個 形聲字聲符標記系統 ", 由中文系研究生與教授的協助, 進行形聲字與其聲符的標記 不過由於此過程需耗費大量時間與人力投入 ( 在 2009/11/10 至 2010/11/23 期間內共有 9593 字由至少三位研究生標記相同聲符 ), 因此是否存在自動判定形聲字聲符的方法, 變成本計畫第一個挑戰 3. 發音相似度比較公式表一般說來, 聲符構件通常與原字的發音相似度高於非聲符構件與原字的發音相似度, 舉例來說, 詁 " 字與其聲符構件 " 古 " 發音相同, 而與其非聲符構件 言 " 發音較不相似, 又如 校 " 字與其聲符構件 " 交 " 發音相近, 而與其非聲符構件 木 " 發音較不相似 因此發音相似度可以做為我們判定一個形聲字聲符的重要依據 為此我們必須有一套漢字發音相似度的計算方法 每一個單獨的漢字雖為單音節發音, 但是就聲韻學上來看可分為聲母 韻母與調性三類 聲母是使用在韻母前面的輔音, 隨著發音部位與發音方法而有所不同, 如表一所示 ; 而韻母則是一個音節中的元音 ( 母音 ), 也是押韻的主要部份 ; 再者由於漢語本身是具備聲調系統的語言, 因此我們在計算一個形聲字與其構件的發音相似度時即可以聲母 韻母 及聲調的相似度做為判斷發音相似度的依據 表一 聲母的發音部位與發音方法 3.1 人工制定聲母和韻母發音相似度如表一所示, 聲母依發音部位的不同, 可分為雙唇 唇齒 舌尖前 舌尖中 舌尖後 舌面前 舌根七種音 ; 若依發音方法的不同, 則可分為塞音 塞擦音 擦音 鼻音 邊音等五類, 加以聲帶的清濁不同, 我們制訂如下之聲母與聲母之間發音相似度 :

77 以最佳化及機率分佈標記形聲字聲符之研究 149 相同的聲母, 相似度訂為 ub=1 國際拼音相同, 但是差一個上標號, 相似度訂為 a=0.9 發音部位相同 ( 同列 ), 相似度訂為 b=0.8 發音方法相同 ( 同行 ), 相似度訂為 c=0.5 其他情形 ( 如不同行不同列或是沒有聲母的情形 ), 相似度訂為 0.1 同時我們也尋求語言學專家的協助, 針對此版本的制訂規則提供意見, 專家們對聲母部分提出以下兩項的建議 : ㄈ與ㄅㄆㄇ的關係合併為同一行, 隸屬唇音一大類 相似度訂為 b ㄗㄘㄙ ㄓㄔㄕ與ㄐㄑㄒ三大類 ( 不包括ㄖ ), 除了同屬塞擦音 擦音外, 在發音上有許多人有彼此相混的現象, 所以相似度定為 d=0.7 表二 韻母種類種類注音符號單韻母 ( 帀 ) ㄧ(yi) ㄨ(wu) ㄩ(yu) ㄚ(a) ㄛ(o) ㄜ(e) ㄝ( ) 複韻母聲隨韻母捲舌韻母結合韻母 ㄞ (ai) ㄟ (ei) ㄠ (ao) ㄡ (ou) ㄢ (an) ㄣ (en) ㄤ (ang) ㄥ (eng) ㄦ (er) ㄧㄚ (ya) 一ㄛ (yo) ㄧㄝ (ye) ㄧㄞ (yai) ㄧㄠ (yau) ㄧㄡ (you) ㄧㄢ (yan) 一ㄣ (yin) ㄧㄤ (yang) ㄧㄥ (ying) ㄨㄚ (wa) ㄨㄛ (wo) ㄨㄞ (wai) ㄨㄟ (wei) ㄨㄢ (wan) ㄨㄣ (wen) ㄨㄤ (wang) ㄨㄥ (weng) ㄩㄝ (yue) ㄩㄢ (yuan) ㄩㄣ (yun) ㄩㄥ (yong) 接著我們制訂韻母與韻母的發音相似度 雖然每個韻母在注音符號表中只是單一個符號, 但是我們可以進一步將韻母分為單韻母 複韻母 聲隨韻母 捲舌韻母與結合韻母五種 其中複韻母包括兩個母音, 聲隨韻母包含韻母後的輔音, 而結合韻母則包含兩個韻母 依據以上分類, 我們制訂韻母發音相似度如下 : 韻母相同則相似度訂為 1 若兩個韻母同為結合韻母, 若兩者有共同尾音則設相似度為 x=0.8, 若兩者有共同第一個音則設相似度為 y=0.5, 否則設相似度為 lb=0.1 例如: ㄧㄚ和ㄨㄚ有共同尾音ㄚ因此相似度為 x ㄨㄤ和ㄨㄢ有相同第一個韻母因此相似度為 y; ㄧㄞ和ㄨㄟ因無共同部分, 設相似度為 lb 若兩個韻母均非結合韻母 ( 也就是屬於單韻母 複韻母 聲隨韻母或捲舌韻母 ), 則相似度以國際拼音決定 若拼音出現相同字母則設相似度為 z=0.5, 否則設相似度為 lb 例如: ㄧ (i) 和ㄞ (ai) 有相同部分 (i) ㄠ(au) 和ㄡ (ou) 間有相同部分 (u) 皆設相似度為 z

78 150 張嘉惠等 若一則為結合韻母, 一則非結合韻母, 則給分方式如下所示 : 若單韻母出現在結合韻母中的後面位置設相似度為 x=0.8, 若出現在前面則設相似度為 y=0.5, 否則相似度為 lb 例如: ㄧ出現在ㄧㄚ前面位置, 則設相似度為 y=0.5; 而ㄞ出現在一ㄞ後面位置, 設相似度則為 x=0.8 其他情形, 相似度訂為 lb 根據上述規則所制訂的聲母與韻母相似度表分別列於附錄 A 及 B 給定以上相似度, 我們定義兩個漢字發音相似度為其聲母相似度 韻母相似度的總和 : Similarity(c1,c2)=Initial(c1,c2)+Vowel(c1,c2) (1) 因此給定一個形聲字, 我們依據漢字構詞資料庫所拆解成的二至三個構件, 分別計算這些構件與原本漢字的發音相似度, 查閱聲母與韻母的相似度比較公式表, 求算聲母與韻母的總和, 取相似度大者構件, 做為聲符的預測 因此形聲字 w 的聲符即可選取與 w 發聲最為相似的構件 c PC( w) = arg max Similarity( w, c) (2) c w 舉例來說, 漢字 校 ( ㄒㄧㄠ 4) 的構件為 木 ( ㄇㄨ 4) 和 交 ( ㄐㄧㄠ 1), 採用相似度比較公式表求算 木 和 校 的分數, 其中聲母不同行不同列, 相似度為 0.1, 而韻母無共同音因此相似度則為 lb, 故相似度總和為 0.1+lb =0.2; 同理 交 和 校 的聲母發音部位相同, 相似度為 b, 且韻母同為結合韻母, 相似度為 1, 故相似度總和為 b+1 =1.8, 因此系統判定 交 為 校 的聲符 若各構件的總和皆相同, 則加入調性進行校正, 選擇與原字調性相同的構件做為預測聲符 舉例而言, 漢字 祖 ( ㄗㄨ 3) 的構件為 示 ( ㄕ 4) 和 且 ( ㄑㄧㄝ 3), 採用相似度比較公式表求算 示 和 祖 的相似度分數為 d+0.1=0.8, 且 和 祖 的總和為 d+lb=0.8, 兩者相似度分數相同, 因此我們再加上調性判別, 預測與 祖 聲調相同的 且 為聲符 3.2 發音相似度最佳化由於前述發音相似度比較公式表是由人工制訂, 這些值是否能有效的做為聲符預測的參數, 還是有更佳的值可以推測形聲字聲符? 另外, 聲母 韻母及聲調三者所占的比重為何? 則是本節所要探討的問題 我們嘗試採用限制型最佳化方法計算聲母和韻母之發音相似度 假設一組已知聲符的形聲字 T, 依照發音相似度比較公式, 我們可以為每一個形聲字 w T 列出 w 的聲符構件與原字發音相似度必須大於非聲符構件與原字發音相似度的限制條件 以前例漢字 校 來說, 其構件為 木 和 交, 而其已知聲符為 交, 因此 Similarity( 木, 校 ) Similarity( 交, 校 ), 也就是 0.1+lb b+1 在我們的問題中, 可以將聲母發音相似度參數 ub, a, b, c, d 以及韻母發音相似度參數 x, y, z, lb, 拿來做為最佳化問題中的變數 由於當限制條件多於變數個數時, 系統可能無解, 因此我們對每個不等式的聲符部份加上一個額外的變數 ε i, 0, 也就是 d+0.1 b+1+ p i i ε i, 再以 ε 做為最小化的目標函數, 確保聲符與原字的發音相似度大於非聲符構件

79 以最佳化及機率分佈標記形聲字聲符之研究 151 與原字的相似度 舉例而言, 若是聲符與原字的相似度小於非聲符構件與原字的相似度, 則 ε i 必須大於 0 才足以讓條件成立, 反之若聲符與原字的相似度已大於非聲符構件與原字的相似度, 則 ε i 在最小化的目標下自然會是 0 因此若有 m 個已知聲符的漢字, 則可化為以下最佳化問題 : min i ε 0.1+ lb <= ub + 1+ ε y <= ub+ 1+ ε2 p i s.t. M b + y <= a ε T abcd,,,, xyzublb,,,,, εi >= 0 其中 p > 0, 代表錯誤聲符對系統的處罰程度的不同 在最極端的情形下, 我們可以針對 21 個聲母之間的相似度設定 22*21/2 個變數 ( 含聲母空的情形 ), 同理也可對 39 個韻母 ( 含空韻 ) 之間的相似度設定 39*38/2 個變數, 再以限制型最佳化的方法來找出對聲符預測最有利的相似度分數 但由於會有將近 1000 個變數, 所花的時間相對也比較長 機率分佈比較法除了前述兩項發音相似度比較公式表與最佳化分析的方法, 我們也從另一個角度觀察漢字的發音, 我們發現某些漢字構件有較強的發音強度, 常常做為聲符, 而屬於部首的構件, 則通常代表字的形意 於是我們假設漢字的發音有可能是由其構件發音強度較高的構件所支配 因此, 如何制定構件的發音強度而又不耗費大量的人力是我們的首要目標 我們觀察一些常見的漢字, 如表三所示 從中不難發現包含構件 包 的漢字的發音不管在聲母 韻母或聲調的表現一致性較高, 而包含構件 火 的漢字則較低, 一致性較高的構件我們也可以說它是發音集中在某幾種發音上, 反之較為分散 集中度較高的構件就很有可能支配著包含此構件的漢字發音, 也就是構件發音強度較強 假設 S 代表某些漢字所形成的集合,ƒ(S) g(s) h(s) 分表示其聲母 韻母及聲調的分佈機率 令 A 表示所有漢字所成的集合, 則 ƒ(a) g(a) h(a) 分別表示漢字的聲母 韻母及聲調的分佈機率 同理對於一個漢字構件 b, 我們可以找出包含 b 的所有漢字 B, 同時求得其聲母 韻母及聲調的分佈機率 ƒ(b) g(b) h(b) 若是 b 發音集中度較高, 則其聲母分佈 ƒ(b) 與 ƒ(a) 就會有較大的差異 因此我們採用 Kullback Leibler divergence 的方法來計算兩個分佈的距離 Kullback Leibler divergence 的公式如下 : P(i) KL (P Q) = P(i)log (4) i Q(i) 因此我們可以計算 KL(ƒ(B) ƒ(a)) 做為構件 b 聲母強度, 同理計算 KL(g(B) g(a)) 做為 b 韻母強度, 以及計算 KL(h(B) h(a)) 做為聲調強度 以下我們以調號以下我們以聲母為例, 計算構件 火 及 包 的聲母強度, 我們從漢字構詞資料庫中找出所有標示注音的字共 A = 14598; 我們同時統計含有構件 火 的字共有 B =259, 含有構件 包 的字共有 C =32, 其聲母分佈如下 : (3)

80 152 張嘉惠等 表三 全部漢字 A 及包含構件 火 B 的聲母分佈狀況 ㄅㄆㄇㄈㄉㄊㄋㄌㄍㄎㄏ 所有漢字 A = f(a) 含構件火 B = f(b) 含構件包 C = f(c) ㄑㄒㄓㄔㄕㄖㄗㄘㄙ空 所有漢字 A = f(a) 含構件火 B = f(b) 含構件包 C = f(c) 將 A 與 B 正規化得 f(a) 與 f(b) 兩機率分佈 最後將 f(b) 及 f(a) 代入 KL-divergence 公式可得構件 火 的聲母強度 同樣方式可計算 包 的聲母強度 當我們要判斷某個漢字的聲符時, 只需要將此漢字的所有構件的聲母 韻母及聲調三種 Kullback Leibler divergence 的值做加總, 那麼擁有最大的加總值的構件便會被我們判定為此漢字的聲符 這個方法的好處在於方法簡單而且不需訓練過程, 後續實驗將可看出這個方法對形聲字聲符判斷相當有效 表四分別列出依聲母 韻母 聲調 KL 值與其構件出現次數 B 的乘積值排序前十名構件 這樣子的排序可以做為聲符學習的參考順序

81 以最佳化及機率分佈標記形聲字聲符之研究 153 表四 漢字構件發音強度表 字碼 聲母 KL 值 字碼 韻母 KL 值 字碼 聲調 KL 值 分 非 皇 莫 分 盧 非 令 令 令 票 會 元 莫 夷 票 屯 希 卑 龍 余 弗 皇 吉 方 包 肖 俞 同 世 實驗以下實驗探討前二節所提出的三種分析方法 : 發音相似度比較公式表 發音相似度最佳化與機率分佈比較法, 分別在自動判別形聲字之聲符的效能為何 實驗中使用的測試資料集, 是取自漢字構形資料庫中有注音標示的漢字, 在 2009/11/10 至 2010/11/23 期間內共有 9593 字為四位中文系研究生共同標記完成 4.1 發音相似度比較法首先我們進行發音相似度比較公式的預測效能 如表五所示, 原始發音相似度方法, 準確率約 88.67%, 其中包含 518 筆無法判別的字 ; 加入聲調的判別, 無法判別的字減少至 291, 準確率約 90.21% 採用專家建議的修正版之發音相似度比較公式表來測試, 準確率提高至 89.39%, 再加入聲調後則為九成零七的準確率 顯示以發音相似度比較公式進行判別聲符, 有一定的效果 也因為考慮聲調可以提升一個多百分點, 因此我們將調的相似度直接納入發音相似度的計算, 將計算式 (1) 改為如下計算式 : Similarity(c1,c2)=Initial(c1,c2)+Vowel(c1,c2)+Tone(c1,c2) (5) 其中聲調相似度依兩個發音的聲調是否相同, 給予 δ 及 0 的相似度 參數 δ 的值則由分析每一個字的 pcdiff 的分佈圖而得 方法如下 : 對於每一個字 w, 我們先用方程式 (1) 計算 w 與其聲符構件 pc w 的發音相似度, 以及 w 與其非聲符構件的發音相似度 ( 若有多個非

82 154 張嘉惠等 聲符構件則取發音相似度值較大者 ), 並定義 pcdiff(w) 為兩者的差 : pcdiff ( w) Similarity( w, pcw) similarity( w, c) c w\ pc = (6) w 表五 發音相似度比較法判別準確率正確錯誤無法判別準確率原始版 (no d) 原始版 + 調 修正版 (d=0.7) 修正版 + 調 整合版 (δ=0.2) 圖一為所有字的 pcdiff 分佈圖 (Histogram) 每一個藍色長條圖 M i 代表 pcdiff 為 i (i=, -0.2, -0.1, 0, 0.1, 0.2,.) 的字的個數 ( 黑色字 ), 同時我們也統計每一藍色長條中有多少個字的聲調與其聲符構件相同但與非聲符構件不同, 我們用 N i 來表示 ( 綠色字 ); 另外我們也統計有多少個字的聲調與其聲符構件不同但與非聲符構件相同, 我們用 L i 來表示 ( 紫色字 ) 這兩部份的字集分別代表的是在採用方程式(5) 來計算發音相似度時,pcdiff 會增加或是減少的字數 因此如果我們採用方程式 (5) 來計算發音相似度, 將會有新增加 N 0 +N N -δ+0.1 可被正確預測的字 ( 橘色轉換 ), 但是同時會有 L 0.1 +L L δ 的字會從正確預測轉為錯誤預測或無法判斷 ( 藍色轉換 ) 因此若 N 0 =125, L 0 =73, N -0.1 =45, N -0.2 =10, L 0.1 =25, L 0.2 =14( 如圖二 ), 當我們設 δ=0.2, 則我們可新增 =170 個可正確判斷的字, 但失去 25+14=39 原可正確判斷的字 整體來說, 我們可多預測 =131 個正確的字, 而無法判別的字則減少 =174 個字 ( =268) 圖一 聲符發音相似度與非聲符發音相似度差分佈圖

83 以最佳化及機率分佈標記形聲字聲符之研究 發音相似度最佳化第二部份實驗主要是了解最佳化方法計算出的聲母與韻母相似度參數在判別漢字聲符之可能性 實驗主要目的在了解需要多少訓練資料筆數才能逹到大約九成的準確率, 以及最佳化方法的學習曲線 由於變數不多, 理論上我們所需要的訓練資料筆數並不需要太多, 但是若是取太少訓練資料, 則預測準確的效果差異則會變大 我們取 p=1 做為目標函數, 隨機抽取 1000 筆資料做為訓練資料, 剩餘 8593 筆資料做為測試資料 每次從訓練資料中隨機抽取 m(=20, 60, 80, 100, 200, 500) 個漢字產生 m 個限制條件 ( 公式 1), 再由線性規劃方法計算出參數值, 並利用這些參數來檢視所有 8593 筆測試漢字聲符判別之準確率 我們反覆十次上述的實驗, 用以繪製盒鬚圖 (Box-Whisker Plot) 得到如圖二之結果 Phonetic Component Prediction Testing Accuracy 100% 95% 90% 85% 80% 75% 70% 89.03% 89.78% 89.98% 89.47% 89.90% 89.50% 88.22% # constraints 圖二 以發音相似度最佳化作為聲符預測準確圖如圖二所示, 我們可以藉由最佳化方法找到與前節所得準確率相當的相似度參數, 甚至於在少數已知聲符的漢字訓練資料, 如 40 或 60 筆時, 即有相當好的結果, 超越修正版 89.39% 的準確率, 所得參數值與人工定義的參數設定方式接近, 如聲母參數 ub>a>b>c,d>0.1 等關係, 韻母參數 1>x>y,z>lb 等關係 ( 如表六所示 ) 事實上聲符預測準確率最高 89.98%, 可由 60 筆已知聲符的形聲字所產生的限制條件求出, 然當訓練資料 ( 限制條件 ) 增加時, 準確率並無上升的情形, 甚至在 500 時, 反而有下降趨勢 ( 最佳測試準確率降至 87% 及 81.5%) 推測主要的原因在於聲符與形聲字發音相似度大於非聲符構件與形聲字發音相似度僅是一個通則, 仍有相當多例外的情形 例如, 冶 洛 債 時 枸 荼等字, 前二者產生 y<=lb+ε 的限制, 中二者則產生 b<=d+ε 的限制, 後二者則產生 z<=lb+ε 的限制條件 在少量的訓練資料時, 這些限制條件的影響可能只是影響一二個變數大小關係, 但是在較多的訓練資料時, 將可能同時影響數個變數, 最終是聲母參數均相同 (ub=a=b=c=d), 韻母參數均相同 (x=y=z=lb) 時, 反而可以讓目標函數的值較小 因此我們看到當訓練資料 500 筆時, 即使是最佳一組測試準確率, 也僅得聲母參數均相同的結果

84 156 張嘉惠等 表六 不同筆訓練資料所得最佳測試準確率及其聲母 韻母參數 訓練資料準確率 90.0% 98.0% 98.0% 96.0% 94.0% 98.0% 97.0% 測試資料準確率 89.0% 89.8% 90.0% 89.5% 90.0% 90.0% 88.0% ub a b c d x y z lb Affected by constraints b<=c or d, y<=z, z<=lb 4.3 機率分佈比較法第三部份的實驗目的在於了解機率分佈比較法對於在判別漢字的聲符之效能 如表七所示, 以聲母分佈 韻母分佈以及聲調分佈個別強度做為聲符預測, 都有一定的準確度 ( 分別為 92.8% 97.5% 及 95.9%), 其中又以韻母的分佈是三者當中最為有效方式 整體來說, 三種分佈一起考量的結果有最佳的效果, 針對 9593 筆形聲字, 其中有 9413 筆正確, 180 筆錯誤, 準確率 98.1% 表七 機率分佈比較法判別準確率 正確錯誤準確率聲 韻 調 聲 + 韻 + 調 為了解判定錯誤發生的可能原因, 我們列出錯誤例子如表八 這些例子顯示, 機率分佈比較法發生錯誤多在在於構件間的強度太過接近, 尤其是兩個構件均為部首的情形, 然而若就發音相似度比較法而言應該可以正確判斷其聲符 因此也可以考慮結合機率分佈比較法與發音相似度比較法, 對於形聲字進行聲符的判斷測試

85 以最佳化及機率分佈標記形聲字聲符之研究 157 表八 機率分佈比較法錯誤例子 5. 結論及未來研究本篇論文主要目的是藉由對形聲字的分析研究, 找出漢字與其聲符構件原字之間的關係 在第一階段我們針對漢字形聲字聲符的標記, 除了採用中文系研究生的人力標記之外, 同時也提出三種自動判別的方式, 用以加速形聲字聲符的標記工作 實驗顯示, 不論是人工制訂的相似度參數或是最佳化方式所得的參數, 預測準確率大約可以做到九成 這是否是本篇論文所使用的最佳化方法的不足, 又或是發音相似度比較法對於聲符預測的極限, 仍尚待進一步的研究 另外一方面, 以每個部件發音的分佈集中與否做為聲符的判斷, 則有高達九成八的準確率, 則是相當有用的資訊 雖然形聲字在聲符標注完成後, 預測聲符的需求即消失不在, 但是透過發音相似度最佳化方法所得的聲母, 韻母相似度參數或許有助於未來漢字字音處理的研究, 同時部件發音強度也可做為漢字教學順序參考, 仍有相當的重要性 未來我們將以持續以挑戰漢字的發音規則知識庫為出發, 除解析漢字發音規則外, 也希望以此項研究為出發, 發展一套漢字學習的順序, 讓使用者可用較少的學習時間, 有效率認識更多漢字 致謝 本論文的完成感謝陳怡如 葉博榮 鍾哲宇 趙婕妤等人的幫助 參考資料 許慎撰, 段玉裁注 (1988) 說文解字注, 台北藝文印書館 莊德明 謝清俊 (2005,1 月 ) 漢字構形資料庫的建置與應用, 漢字與全球化國際學術研討會, 台北 莊德明 鄧賢瑛 (2008,8 月 ) 文字學入口網站的規畫, 第四屆中國文字學國際學術研討會, 山東煙台

86 158 張嘉惠等 附錄 A: 聲母相似度比較表

87 附錄 B: 韻母相似度比較表 以最佳化及機率分佈標記形聲字聲符之研究 159

88 160 張嘉惠等

89 Aims: The Association for Computational Linguistics and Chinese Language Processing (new members are welcomed) 1. To conduct research in computational linguistics. 2. To promote the utilization and development of computational linguistics. 3. To encourage research in and development of the field of Chinese computational linguistics both domestically and internationally. 4. To maintain contact with international groups who have similar goals and to cultivate academic exchange. Activities: 1. Holding the Republic of China Computational Linguistics Conference (ROCLING) annually. 2. Facilitating and promoting academic research, seminars, training, discussions, comparative evaluations and other activities related to computational linguistics. 3. Collecting information and materials on recent developments in the field of computational linguistics, domestically and internationally. 4. Publishing pertinent journals, proceedings and newsletters. 5. Setting of the Chinese-language technical terminology and symbols related to computational linguistics. 6. Maintaining contact with international computational linguistics academic organizations. 7. Dealing with various other matters related to the development of computational linguistics. To Register: Please send application to: The Association for Computational Linguistics and Chinese Language Processing Institute of Information Science, Academia Sinica 128, Sec. 2, Academy Rd., Nankang, Taipei 11529, Taiwan, R.O.C. payment: Credit cards(please fill in the order form), cheque, or money orders. Annual Fees: regular/overseas member: NT$ 1,000 (US$50.-) group membership: NT$20,000 (US$1,000.-) life member:ten times the annual fee for regular/ group/ overseas members Contact: Address: The Association for Computational Linguistics and Chinese Language Processing Institute of Information Science, Academia Sinica 128, Sec. 2, Academy Rd., Nankang, Taipei 11529, Taiwan, R.O.C. Tel.: ext Fax: Web Site: Please address all correspondence to Miss Qi Huang, or Miss Abby Ho

90 The Association for Computational Linguistics and Chinese Language Processing Membership Application Form Member ID#: Name: Date of Birth: Country of Residence: Province/State: Passport No.: Sex: Education(highest degree obtained): Work Experience: Present Occupation: Address: Add: Tel. No: Fax No: Membership Category: Regular Member Date: / / (Y-M-D) Life Member Applicant's Signature: Remarks: Please indicated clearly in which membership category you wish to register, according to the following scale of annual membership dues: Regular Member : US$ 50.- (NT$ 1,000) Life Member : US$500.-(NT$10,000) Please feel free to make copies of this application for others to use. Committee Assessment:

91 中華民國計算語言學學會 宗旨 : ( 一 ) 從事計算語言學之研究 ( 二 ) 推行計算語言學之應用與發展 ( 三 ) 促進國內外中文計算語言學之研究與發展 ( 四 ) 聯繫國際有關組織並推動學術交流活動項目 : ( 一 ) 定期舉辦中華民國計算語言學學術會議 (Rocling) ( 二 ) 舉行有關計算語言學之學術研究講習 訓練 討論 觀摩等活動項目 ( 三 ) 收集國內外有關計算語言學知識之圖書及最新發展之資料 ( 四 ) 發行有關之學術刊物, 論文集及通訊 ( 五 ) 研定有關計算語言學專用名稱術語及符號 ( 六 ) 與國際計算語言學學術機構聯繫交流 ( 七 ) 其他有關計算語言發展事項報名方式 : 1. 入會申請書 : 請至本會網頁下載入會申請表, 填妥後郵寄或 至本會 2. 繳交會費 : 劃撥 : 帳號 : , 戶名 : 中華民國計算語言學學會信用卡 : 請至本會網頁下載信用卡付款單年費 : 終身會員 : 10,000.- (US$ 500.-) 個人會員 : 1,000.- (US$ 50.-) 學生會員 : ( 限國內學生 ) 團體會員 : 20,000.- (US$ 1,000.-) 連絡處 : 地址 : 台北市 115 南港區研究院路二段 128 號中研院資訊所 ( 轉 ) 電話 :(02) ext.1502 傳真 :(02) aclclp@hp.iis.sinica.edu.tw 網址 : 連絡人 : 黃琪小姐 何婉如小姐

92 中華民國計算語言學學會個人會員入會申請書 會員類別 終身 個人 學生會員編號 ( 由本會填寫 ) 姓名性別 出生日期年月日 身分證號碼 現職學歷 通訊地址 戶籍地址 電話 申請人 : ( 簽章 ) 中華民國年月日 審查結果 : 1. 年費 : 終身會員 : 10,000.- 個人會員 : 1,000.- 學生會員 : 500.-( 限國內學生 ) 團體會員 : 20, 連絡處 : 地址 : 台北市南港區研究院路二段 128 號中研院資訊所 ( 轉 ) 電話 :(02) ext.1502 傳真 :(02) aclclp@hp.iis.sinica.edu.tw 網址 : 連絡人 : 黃琪小姐 何婉如小姐 3. 本表可自行影印

93 The Association for Computational Linguistics and Chinese Language Processing (ACLCLP) PAYMENT FORM Name : (Please print) Date: Please debit my credit card as follows: US$ VISA CARD MASTER CARD JCB CARD Issue Bank: Card No.: Exp. Date: 3-digit code: (on the back card, inside the signature area, the last three digits) CARD HOLDER SIGNATURE : Tel.: Add: PAYMENT FOR US$ Computational Linguistics & Chinese Languages Processing (CLCLP) Quantity Wanted: US$ US$ US$ US$ Publications: Text Corpora: Speech Corpora: Others: US$ Life Member Fee New Member Renew US$ = Total Fax : or Mail this form to : ACLCLP Institute of Information Science, Academia Sinica R502, 128, Sec.2, Academia Rd., Nankang, Taipei 115, Taiwan aclclp@hp.iis.sinica.edu.tw Website:

94 中華民國計算語言學學會信用卡付款單 姓名 : ( 請以正楷書寫 ) 日期 :: 卡別 : VISA CARD MASTER CARD JCB CARD 發卡銀行 : 卡號 : 有效日期 : 卡片後三碼 : ( 卡片背面簽名欄上數字後三碼 ) 持卡人簽名 : ( 簽名方式請與信用卡背面相同 ) 通訊地址 : 聯絡電話 : 備註 : 為順利取得信用卡授權, 請提供與發卡銀行相同之聯絡資料 付款內容及金額 : NT$ 中文計算語言學期刊 (IJCLCLP) NT$ 中研院詞庫小組技術報告 NT$ 中文 ( 新聞 ) 語料庫 NT$ 平衡語料庫 NT$ 中文詞庫八萬目 NT$ 中文句結構樹資料庫 NT$ 平衡語料庫詞集及詞頻統計 NT$ 中英雙語詞網 NT$ 中英雙語知識庫 NT$ 語音資料庫 NT$ 會員年費 續會 新會員 終身會員 NT$ 其他 : NT$ = 合計 填妥後請傳真至 或郵寄至 : 115 台北市南港區研究院路 2 段 128 號中研院資訊所 ( 轉 ) 中華民國計算語言學學會收 aclclp@hp.iis.sinica.edu.tw Website:

95 Publications of the Association for Computational Linguistics and Chinese Language Processing Surface AIR (US&EURP) AIR (ASIA) VOLUME AMOUNT no.92-01, no ( 合訂本 ) ICG 中的論旨角色與 A Conceptual Structure for Parsing Mandarin -- Its Frame and General Applications-- US$ 9 US$ 19 US$15 no V-N 複合名詞討論篇 & V-R 複合動詞討論篇 no 新聞語料庫字頻統計表 no 新聞語料庫詞頻統計表 no 新聞常用動詞詞頻與分類 no 中文詞類分析 no 現代漢語中的法相詞 no 中文書面語頻率詞典 ( 新聞語料詞頻統計 ) no 古漢語字頻表 no 注音檢索現代漢語字頻表 no.95-02/98-04 中央研究院平衡語料庫的內容與說明 no 訊息為本的格位語法與其剖析方法 no 搜 文解字- 中文詞界研究與資訊用分詞標準 no 古漢語詞頻表 ( 甲 ) no 論語詞頻表 no 詞頻詞典 no Accumulated Word Frequency in CKIP Corpus no 自然語言處理及計算語言學相關術語中英對譯表 no 現代漢語口語對話語料庫標註系統說明 Computational Linguistics & Chinese Languages Processing (One year) (Back issues of IJCLCLP: US$ 20 per copy) Readings in Chinese Language Processing TOTAL 10% member discount: Total Due: OVERSEAS USE ONLY PAYMENT: Credit Card ( Preferred ) Money Order or Check payable to "The Association for Computation Linguistics and Chinese Language Processing " or 中華民國計算語言學學會 aclclp@hp.iis.sinica.edu.tw Name (please print): Fax: Signature: Address:

96 中華民國計算語言學學會相關出版品價格表及訂購單 編號 書目 會員 非會員 冊數 金額 1. no.92-01, no ( 合訂本 ) ICG 中的論旨角色與 A conceptual Structure for Parsing Mandarin--its Frame and General Applications-- NT$ 80 NT$ no.92-02, no ( 合訂本 ) V-N 複合名詞討論篇與 V-R 複合動詞討論篇 no 新聞語料庫字頻統計表 no 新聞語料庫詞頻統計表 no 新聞常用動詞詞頻與分類 no 中文詞類分析 no 現代漢語中的法相詞 no 中文書面語頻率詞典 ( 新聞語料詞頻統計 ) no 古漢語字頻表 no 注音檢索現代漢語字頻表 no.95-02/98-04 中央研究院平衡語料庫的內容與說明 no 訊息為本的格位語法與其剖析方法 no 搜 文解字 - 中文詞界研究與資訊用分詞標準 no 古漢語詞頻表 ( 甲 ) no 論語詞頻表 no 詞頻詞典 no Accumulated Word Frequency in CKIP Corpus no 自然語言處理及計算語言學相關術語中英對譯表 no 現代漢語口語對話語料庫標註系統說明 論文集 COLING 2002 紙本 論文集 COLING 2002 光碟片 論文集 COLING 2002 Workshop 光碟片 論文集 ISCSLP 2002 光碟片 交談系統暨語境分析研討會講義 ( 中華民國計算語言學學會 1997 第四季學術活動 ) 中文計算語言學期刊 ( 一年四期 ) 年份 : ( 過期期刊每本售價 500 元 ) --- 2, Readings of Chinese Language Processing 剖析策略與機器翻譯 此價格表僅限國內 ( 台灣地區 ) 使用 劃撥帳戶 : 中華民國計算語言學學會 劃撥帳號 : 聯絡電話 :(02) 轉 1502 聯絡人 : 黃琪小姐 何婉如小姐 aclclp@hp.iis.sinica.edu.tw 訂購者 : 收據抬頭 : 地 址 : 電 話 : 合計

97

98

Microsoft Word - TIP006SCH Uni-edit Writing Tip - Presentperfecttenseandpasttenseinyourintroduction readytopublish

Microsoft Word - TIP006SCH Uni-edit Writing Tip - Presentperfecttenseandpasttenseinyourintroduction readytopublish 我 难 度 : 高 级 对 们 现 不 在 知 仍 道 有 听 影 过 响 多 少 那 次 么 : 研 英 究 过 文 论 去 写 文 时 作 的 表 技 引 示 巧 言 事 : 部 情 引 分 发 言 该 生 使 在 中 用 过 去, 而 现 在 完 成 时 仅 表 示 事 情 发 生 在 过 去, 并 的 哪 现 种 在 时 完 态 成 呢 时? 和 难 过 道 去 不 时 相 关? 是 所 有

More information

國立中山大學學位論文典藏.PDF

國立中山大學學位論文典藏.PDF I II III The Study of Factors to the Failure or Success of Applying to Holding International Sport Games Abstract For years, holding international sport games has been Taiwan s goal and we are on the way

More information

國家圖書館典藏電子全文

國家圖書館典藏電子全文 i ii Abstract The most important task in human resource management is to encourage and help employees to develop their potential so that they can fully contribute to the organization s goals. The main

More information

2/80 2

2/80 2 2/80 2 3/80 3 DSP2400 is a high performance Digital Signal Processor (DSP) designed and developed by author s laboratory. It is designed for multimedia and wireless application. To develop application

More information

WTO

WTO 10384 200015128 UDC Exploration on Design of CIB s Human Resources System in the New Stage (MBA) 2004 2004 2 3 2004 3 2 0 0 4 2 WTO Abstract Abstract With the rapid development of the high and new technique

More information

Microsoft PowerPoint _代工實例-1

Microsoft PowerPoint _代工實例-1 4302 動態光散射儀 (Dynamic Light Scattering) 代工實例與結果解析 生醫暨非破壞性分析團隊 2016.10 updated Which Size to Measure? Diameter Many techniques make the useful and convenient assumption that every particle is a sphere. The

More information

東莞工商總會劉百樂中學

東莞工商總會劉百樂中學 /2015/ 頁 (2015 年 版 ) 目 錄 : 中 文 1 English Language 2-3 數 學 4-5 通 識 教 育 6 物 理 7 化 學 8 生 物 9 組 合 科 學 ( 化 學 ) 10 組 合 科 學 ( 生 物 ) 11 企 業 會 計 及 財 務 概 論 12 中 國 歷 史 13 歷 史 14 地 理 15 經 濟 16 資 訊 及 通 訊 科 技 17 視 覺

More information

2015 Chinese FL Written examination

2015 Chinese FL Written examination Victorian Certificate of Education 2015 SUPERVISOR TO ATTACH PROCESSING LABEL HERE Letter STUDENT NUMBER CHINESE FIRST LANGUAGE Written examination Monday 16 November 2015 Reading time: 11.45 am to 12.00

More information

4. 每 组 学 生 将 写 有 习 语 和 含 义 的 两 组 卡 片 分 别 洗 牌, 将 顺 序 打 乱, 然 后 将 两 组 卡 片 反 面 朝 上 置 于 课 桌 上 5. 学 生 依 次 从 两 组 卡 片 中 各 抽 取 一 张, 展 示 给 小 组 成 员, 并 大 声 朗 读 卡

4. 每 组 学 生 将 写 有 习 语 和 含 义 的 两 组 卡 片 分 别 洗 牌, 将 顺 序 打 乱, 然 后 将 两 组 卡 片 反 面 朝 上 置 于 课 桌 上 5. 学 生 依 次 从 两 组 卡 片 中 各 抽 取 一 张, 展 示 给 小 组 成 员, 并 大 声 朗 读 卡 Tips of the Week 课 堂 上 的 英 语 习 语 教 学 ( 二 ) 2015-04-19 吴 倩 MarriottCHEI 大 家 好! 欢 迎 来 到 Tips of the Week! 这 周 我 想 和 老 师 们 分 享 另 外 两 个 课 堂 上 可 以 开 展 的 英 语 习 语 教 学 活 动 其 中 一 个 活 动 是 一 个 充 满 趣 味 的 游 戏, 另 外

More information

Microsoft Word doc

Microsoft Word doc 中 考 英 语 科 考 试 标 准 及 试 卷 结 构 技 术 指 标 构 想 1 王 后 雄 童 祥 林 ( 华 中 师 范 大 学 考 试 研 究 院, 武 汉,430079, 湖 北 ) 提 要 : 本 文 从 结 构 模 式 内 容 要 素 能 力 要 素 题 型 要 素 难 度 要 素 分 数 要 素 时 限 要 素 等 方 面 细 致 分 析 了 中 考 英 语 科 试 卷 结 构 的

More information

32 戲劇學刊 A Study of Beijing Opera s Jing Actors and Their Vocal Accents in the Early Twentieth Century Using Two Operas, Muhuguan and Yuguoyuan, as Exa

32 戲劇學刊 A Study of Beijing Opera s Jing Actors and Their Vocal Accents in the Early Twentieth Century Using Two Operas, Muhuguan and Yuguoyuan, as Exa 李元皓 二十世紀初期京劇淨行演員及其唱腔研究 以 牧虎關 御果園 為例 二十世紀初期京劇淨行演員及其唱腔研 究 以 牧虎關 御果園 為例* 李元皓** 中文摘要 形成於十九世紀的京劇 在二十世紀初時 發展的勢頭風潮臻於極盛 新興 的唱片錄音科技也於十九世紀末抵達中國 留下了一批珍貴的老唱片 以不同於 書面文字的方式 記錄著京劇有聲的過往 老唱片所能記錄的雖然只有聲音 然 而聲音 亦即 唱腔 正是戲曲藝術的核心之一

More information

致 谢 本 人 自 2008 年 6 月 从 上 海 外 国 语 大 学 毕 业 之 后, 于 2010 年 3 月 再 次 进 入 上 外, 非 常 有 幸 成 为 汉 语 国 际 教 育 专 业 的 研 究 生 回 顾 三 年 以 来 的 学 习 和 生 活, 顿 时 感 觉 这 段 时 间 也

致 谢 本 人 自 2008 年 6 月 从 上 海 外 国 语 大 学 毕 业 之 后, 于 2010 年 3 月 再 次 进 入 上 外, 非 常 有 幸 成 为 汉 语 国 际 教 育 专 业 的 研 究 生 回 顾 三 年 以 来 的 学 习 和 生 活, 顿 时 感 觉 这 段 时 间 也 精 英 汉 语 和 新 实 用 汉 语 课 本 的 对 比 研 究 The Comparative Study of Jing Ying Chinese and The New Practical Chinese Textbook 专 业 : 届 别 : 姓 名 : 导 师 : 汉 语 国 际 教 育 2013 届 王 泉 玲 杨 金 华 1 致 谢 本 人 自 2008 年 6 月 从 上 海 外

More information

國立臺南大學數位論文典藏.pdf

國立臺南大學數位論文典藏.pdf 1974~2003 Xiang Yang and the Research upon His Modern Poems: 1974~2003 1974~2003 Xiang Yang and the Research upon His Modern Poems: 1974~2003 by StudentSu-Chen Lee A Thesis submitted in partial fulfillment

More information

謝 辭 能 夠 完 成 這 本 論 文, 首 先 當 然 要 感 謝 的 是 我 的 指 導 教 授, 謝 林 德 老 師 這 段 時 間 老 師 不 厭 其 煩 的 幫 我 檢 視 論 文, 每 次 與 老 師 討 論 後 總 是 收 穫 很 多, 在 臺 灣 求 的 這 段 期 間 深 深 地

謝 辭 能 夠 完 成 這 本 論 文, 首 先 當 然 要 感 謝 的 是 我 的 指 導 教 授, 謝 林 德 老 師 這 段 時 間 老 師 不 厭 其 煩 的 幫 我 檢 視 論 文, 每 次 與 老 師 討 論 後 總 是 收 穫 很 多, 在 臺 灣 求 的 這 段 期 間 深 深 地 文 教 碩 博 士 位 程 碩 士 位 論 文 指 導 教 授 : 謝 林 德 博 士 Dr. Dennis Schilling 華 同 形 漢 字 詞 之 比 較 及 教 建 議 以 台 灣 八 千 詞 及 韓 漢 字 語 辭 典 分 析 為 例 Semantic and Pragmatic Features of Chinese and Korean Homographic Words with

More information

Microsoft Word - Final Exam Review Packet.docx

Microsoft Word - Final Exam Review Packet.docx Do you know these words?... 3.1 3.5 Can you do the following?... Ask for and say the date. Use the adverbial of time correctly. Use Use to ask a tag question. Form a yes/no question with the verb / not

More information

2017 CCAFL Chinese in Context

2017 CCAFL Chinese in Context Student/Registration Number Centre Number 2017 PUBLIC EXAMINATION Chinese in Context Reading Time: 10 minutes Working Time: 2 hours and 30 minutes You have 10 minutes to read all the papers and to familiarise

More information

摘 要 張 捷 明 是 台 灣 當 代 重 要 的 客 語 兒 童 文 學 作 家, 他 的 作 品 記 錄 著 客 家 人 的 思 想 文 化 與 觀 念, 也 曾 榮 獲 多 項 文 學 大 獎 的 肯 定, 對 台 灣 這 塊 土 地 上 的 客 家 人 有 著 深 厚 的 情 感 張 氏 於

摘 要 張 捷 明 是 台 灣 當 代 重 要 的 客 語 兒 童 文 學 作 家, 他 的 作 品 記 錄 著 客 家 人 的 思 想 文 化 與 觀 念, 也 曾 榮 獲 多 項 文 學 大 獎 的 肯 定, 對 台 灣 這 塊 土 地 上 的 客 家 人 有 著 深 厚 的 情 感 張 氏 於 玄 奘 大 學 中 國 語 文 學 系 碩 士 論 文 客 家 安 徒 生 張 捷 明 童 話 研 究 指 導 教 授 : 羅 宗 濤 博 士 研 究 生 : 黃 春 芳 撰 中 華 民 國 一 0 二 年 六 月 摘 要 張 捷 明 是 台 灣 當 代 重 要 的 客 語 兒 童 文 學 作 家, 他 的 作 品 記 錄 著 客 家 人 的 思 想 文 化 與 觀 念, 也 曾 榮 獲 多 項 文

More information

Chn 116 Neh.d.01.nis

Chn 116 Neh.d.01.nis 31 尼 希 米 书 尼 希 米 的 祷 告 以 下 是 哈 迦 利 亚 的 儿 子 尼 希 米 所 1 说 的 话 亚 达 薛 西 王 朝 二 十 年 基 斯 流 月 *, 我 住 在 京 城 书 珊 城 里 2 我 的 兄 弟 哈 拿 尼 和 其 他 一 些 人 从 犹 大 来 到 书 珊 城 我 向 他 们 打 听 那 些 劫 后 幸 存 的 犹 太 人 家 族 和 耶 路 撒 冷 的 情 形

More information

天 主 教 輔 仁 大 學 社 會 學 系 學 士 論 文 小 別 勝 新 婚? 久 別 要 離 婚? 影 響 遠 距 家 庭 婚 姻 感 情 因 素 之 探 討 Separate marital relations are getting better or getting worse? -Exp

天 主 教 輔 仁 大 學 社 會 學 系 學 士 論 文 小 別 勝 新 婚? 久 別 要 離 婚? 影 響 遠 距 家 庭 婚 姻 感 情 因 素 之 探 討 Separate marital relations are getting better or getting worse? -Exp 天 主 教 輔 仁 大 學 社 會 學 系 學 士 論 文 小 別 勝 新 婚? 久 別 要 離 婚? 影 響 遠 距 家 庭 婚 姻 感 情 因 素 之 探 討 Separate marital relations are getting better or getting worse? -Explore the impact of emotional factors couples do not

More information

Shanghai International Studies University THE STUDY AND PRACTICE OF SITUATIONAL LANGUAGE TEACHING OF ADVERB AT BEGINNING AND INTERMEDIATE LEVEL A Thes

Shanghai International Studies University THE STUDY AND PRACTICE OF SITUATIONAL LANGUAGE TEACHING OF ADVERB AT BEGINNING AND INTERMEDIATE LEVEL A Thes 上 海 外 国 语 大 学 硕 士 学 位 论 文 对 外 汉 语 初 中 级 副 词 情 境 教 学 研 究 与 实 践 院 系 : 国 际 文 化 交 流 学 院 学 科 专 业 : 汉 语 国 际 教 育 姓 名 : 顾 妍 指 导 教 师 : 缪 俊 2016 年 5 月 Shanghai International Studies University THE STUDY AND PRACTICE

More information

A Study on Grading and Sequencing of Senses of Grade-A Polysemous Adjectives in A Syllabus of Graded Vocabulary for Chinese Proficiency 2002 I II Abstract ublished in 1992, A Syllabus of Graded Vocabulary

More information

附件1:

附件1: 附 件 1: 全 国 优 秀 教 育 硕 士 专 业 学 位 论 文 推 荐 表 单 位 名 称 : 西 南 大 学 论 文 题 目 填 表 日 期 :2014 年 4 月 30 日 数 学 小 组 合 作 学 习 的 课 堂 管 理 攻 硕 期 间 及 获 得 硕 士 学 位 后 一 年 内 获 得 与 硕 士 学 位 论 文 有 关 的 成 果 作 者 姓 名 论 文 答 辩 日 期 学 科 专

More information

硕 士 学 位 论 文 论 文 题 目 : 北 岛 诗 歌 创 作 的 双 重 困 境 专 业 名 称 : 中 国 现 当 代 文 学 研 究 方 向 : 中 国 新 诗 研 究 论 文 作 者 : 奚 荣 荣 指 导 老 师 : 姜 玉 琴 2014 年 12 月

硕 士 学 位 论 文 论 文 题 目 : 北 岛 诗 歌 创 作 的 双 重 困 境 专 业 名 称 : 中 国 现 当 代 文 学 研 究 方 向 : 中 国 新 诗 研 究 论 文 作 者 : 奚 荣 荣 指 导 老 师 : 姜 玉 琴 2014 年 12 月 硕 士 学 位 论 文 论 文 题 目 : 北 岛 诗 歌 创 作 的 双 重 困 境 专 业 名 称 : 中 国 现 当 代 文 学 研 究 方 向 : 中 国 新 诗 研 究 论 文 作 者 : 奚 荣 荣 指 导 老 师 : 姜 玉 琴 2014 年 12 月 致 谢 文 学 是 我 们 人 类 宝 贵 的 精 神 财 富 两 年 半 的 硕 士 学 习 让 我 进 一 步 接 近 文 学,

More information

20

20 37 92 19 40 19 20 21 1 7 22 1/5 6/30 5/3030 23 24 25 26 1 2 27 1 2 28 29 30 5 8 8 3 31 32 33 34 35 36 37 38 39 A Study Investigating Elementary School Students Concept of the Unit in Fraction in Northern

More information

UDC The Policy Risk and Prevention in Chinese Securities Market

UDC The Policy Risk and Prevention in Chinese Securities Market 10384 200106013 UDC The Policy Risk and Prevention in Chinese Securities Market 2004 5 2004 2004 2004 5 : Abstract Many scholars have discussed the question about the influence of the policy on Chinese

More information

南華大學數位論文

南華大學數位論文 南 華 大 學 ( 文 學 所 ) 碩 士 論 文 論 文 題 目 ( 陳 千 武 小 說 活 著 回 來 及 其 相 關 事 例 研 究 ) 論 文 題 目 (Chen Chien Wu Return Alive And Some Research About It) 研 究 生 : 朱 妍 淩 指 導 教 授 : 林 葉 連 中 華 民 國 一 0 一 年 6 月 8 日 陳 千 武 小 說

More information

Microsoft Word - 第四組心得.doc

Microsoft Word - 第四組心得.doc 徐 婉 真 這 四 天 的 綠 島 人 權 體 驗 營 令 我 印 象 深 刻, 尤 其 第 三 天 晚 上 吳 豪 人 教 授 的 那 堂 課, 他 讓 我 聽 到 不 同 於 以 往 的 正 義 之 聲 轉 型 正 義, 透 過 他 幽 默 熱 情 的 語 調 激 起 了 我 對 政 治 的 興 趣, 願 意 在 未 來 多 關 心 社 會 多 了 解 政 治 第 一 天 抵 達 綠 島 不 久,

More information

ENGG1410-F Tutorial 6

ENGG1410-F Tutorial 6 Jianwen Zhao Department of Computer Science and Engineering The Chinese University of Hong Kong 1/16 Problem 1. Matrix Diagonalization Diagonalize the following matrix: A = [ ] 1 2 4 3 2/16 Solution The

More information

Construction of Chinese pediatric standard database A Dissertation Submitted for the Master s Degree Candidate:linan Adviser:Prof. Han Xinmin Nanjing

Construction of Chinese pediatric standard database A Dissertation Submitted for the Master s Degree Candidate:linan Adviser:Prof. Han Xinmin Nanjing 密 级 : 公 开 学 号 :20081209 硕 士 学 位 论 文 中 医 儿 科 标 准 数 据 库 建 设 研 究 研 究 生 李 楠 指 导 教 师 学 科 专 业 所 在 学 院 毕 业 时 间 韩 新 民 教 授 中 医 儿 科 学 第 一 临 床 医 学 院 2011 年 06 月 Construction of Chinese pediatric standard database

More information

hks298cover&back

hks298cover&back 2957 6364 2377 3300 2302 1087 www.scout.org.hk scoutcraft@scout.org.hk 2675 0011 5,500 Service and Scouting Recently, I had an opportunity to learn more about current state of service in Hong Kong

More information

Preface This guide is intended to standardize the use of the WeChat brand and ensure the brand's integrity and consistency. The guide applies to all d

Preface This guide is intended to standardize the use of the WeChat brand and ensure the brand's integrity and consistency. The guide applies to all d WeChat Search Visual Identity Guidelines WEDESIGN 2018. 04 Preface This guide is intended to standardize the use of the WeChat brand and ensure the brand's integrity and consistency. The guide applies

More information

A VALIDATION STUDY OF THE ACHIEVEMENT TEST OF TEACHING CHINESE AS THE SECOND LANGUAGE by Chen Wei A Thesis Submitted to the Graduate School and Colleg

A VALIDATION STUDY OF THE ACHIEVEMENT TEST OF TEACHING CHINESE AS THE SECOND LANGUAGE by Chen Wei A Thesis Submitted to the Graduate School and Colleg 上 海 外 国 语 大 学 SHANGHAI INTERNATIONAL STUDIES UNIVERSITY 硕 士 学 位 论 文 MASTER DISSERTATION 学 院 国 际 文 化 交 流 学 院 专 业 汉 语 国 际 教 育 硕 士 题 目 届 别 2010 届 学 生 陈 炜 导 师 张 艳 莉 副 教 授 日 期 2010 年 4 月 A VALIDATION STUDY

More information

Microsoft Word - ChineseSATII .doc

Microsoft Word - ChineseSATII .doc 中 文 SAT II 冯 瑶 一 什 么 是 SAT II 中 文 (SAT Subject Test in Chinese with Listening)? SAT Subject Test 是 美 国 大 学 理 事 会 (College Board) 为 美 国 高 中 生 举 办 的 全 国 性 专 科 标 准 测 试 考 生 的 成 绩 是 美 国 大 学 录 取 新 生 的 重 要 依

More information

穨control.PDF

穨control.PDF TCP congestion control yhmiu Outline Congestion control algorithms Purpose of RFC2581 Purpose of RFC2582 TCP SS-DR 1998 TCP Extensions RFC1072 1988 SACK RFC2018 1996 FACK 1996 Rate-Halving 1997 OldTahoe

More information

<4D6963726F736F667420576F7264202D2032303130C4EAC0EDB9A4C0E04142BCB6D4C4B6C1C5D0B6CFC0FDCCE2BEABD1A15F325F2E646F63>

<4D6963726F736F667420576F7264202D2032303130C4EAC0EDB9A4C0E04142BCB6D4C4B6C1C5D0B6CFC0FDCCE2BEABD1A15F325F2E646F63> 2010 年 理 工 类 AB 级 阅 读 判 断 例 题 精 选 (2) Computer mouse How does the mouse work? We have to start at the bottom, so think upside down for now. It all starts with mouse ball. As the mouse ball in the bottom

More information

341-367

341-367 BIBLID 0254-4466(2000)18:2 pp. 341-367 18 2 89 12 341 342 1 2 3 1 19 1983 2 3 1996. 10 1218 343 4 5 6 7 11 8 28 9 10 11 12 13 14 4 3 1808 5 3 1349 6 3 343 7 3 1292 8 9 1996 4 988 10 3 187 11 3 1506 12

More information

Microsoft Word - 論文封面-980103修.doc

Microsoft Word - 論文封面-980103修.doc 淡 江 大 學 中 國 文 學 學 系 碩 士 在 職 專 班 碩 士 論 文 指 導 教 授 : 呂 正 惠 蘇 敏 逸 博 士 博 士 倚 天 屠 龍 記 愛 情 敘 事 之 研 究 研 究 生 : 陳 麗 淑 撰 中 華 民 國 98 年 1 月 淡 江 大 學 研 究 生 中 文 論 文 提 要 論 文 名 稱 : 倚 天 屠 龍 記 愛 情 敘 事 之 研 究 頁 數 :128 校 系 (

More information

Microsoft PowerPoint - NCBA_Cattlemens_College_Darrh_B

Microsoft PowerPoint - NCBA_Cattlemens_College_Darrh_B Introduction to Genetics Darrh Bullock University of Kentucky The Model Trait = Genetics + Environment Genetics Additive Predictable effects that get passed from generation to generation Non-Additive Primarily

More information

國立中山大學學位論文典藏.PDF

國立中山大學學位論文典藏.PDF 中 國 文 學 系 國 立 中 山 大 學, 碩 士 論 文 國 立 中 山 大 學 中 國 文 學 系 碩 士 論 文 Department of Chinese Literature 肉 蒲 團 研 究 National Sun Yat-sen University Master Thesis 肉 蒲 團 研 究 The Research of Rou Pu Tuan 研 究 生 : 林 欣 穎

More information

Microsoft Word - 01李惠玲ok.doc

Microsoft Word - 01李惠玲ok.doc 康 寧 學 報 11:1-20(2009) 1 數 位 學 習 於 護 理 技 術 課 程 之 運 用 與 評 值 * 李 惠 玲 ** 高 清 華 *** 呂 莉 婷 摘 要 背 景 : 網 路 科 技 在 教 育 的 使 用 已 成 為 一 種 有 利 的 教 學 輔 助 工 具 網 路 教 學 的 特 性, 在 使 學 習 可 不 分 時 間 與 空 間 不 同 進 度 把 握 即 時 性 資

More information

10384 200115009 UDC Management Buy-outs MBO MBO MBO 2002 MBO MBO MBO MBO 000527 MBO MBO MBO MBO MBO MBO MBO MBO MBO MBO MBO Q MBO MBO MBO Abstract Its related empirical study demonstrates a remarkable

More information

University of Science and Technology of China A dissertation for master s degree Research of e-learning style for public servants under the context of

University of Science and Technology of China A dissertation for master s degree Research of e-learning style for public servants under the context of 中 国 科 学 技 术 大 学 硕 士 学 位 论 文 新 媒 体 环 境 下 公 务 员 在 线 培 训 模 式 研 究 作 者 姓 名 : 学 科 专 业 : 导 师 姓 名 : 完 成 时 间 : 潘 琳 数 字 媒 体 周 荣 庭 教 授 二 一 二 年 五 月 University of Science and Technology of China A dissertation for

More information

2015年4月11日雅思阅读预测机经(新东方版)

2015年4月11日雅思阅读预测机经(新东方版) 剑 桥 雅 思 10 第 一 时 间 解 析 阅 读 部 分 1 剑 桥 雅 思 10 整 体 内 容 统 计 2 剑 桥 雅 思 10 话 题 类 型 从 以 上 统 计 可 以 看 出, 雅 思 阅 读 的 考 试 话 题 一 直 广 泛 多 样 而 题 型 则 稳 中 有 变 以 剑 桥 10 的 test 4 为 例 出 现 的 三 篇 文 章 分 别 是 自 然 类, 心 理 研 究 类,

More information

<4D6963726F736F667420506F776572506F696E74202D20B5DAD2BBD5C228B4F2D3A1B0E6292E707074205BBCE6C8DDC4A3CABD5D>

<4D6963726F736F667420506F776572506F696E74202D20B5DAD2BBD5C228B4F2D3A1B0E6292E707074205BBCE6C8DDC4A3CABD5D> Homeworks ( 第 三 版 ):.4 (,, 3).5 (, 3).6. (, 3, 5). (, 4).4.6.7 (,3).9 (, 3, 5) Chapter. Number systems and codes 第 一 章. 数 制 与 编 码 . Overview 概 述 Information is of digital forms in a digital system, and

More information

Microsoft PowerPoint - STU_EC_Ch08.ppt

Microsoft PowerPoint - STU_EC_Ch08.ppt 樹德科技大學資訊工程系 Chapter 8: Counters Shi-Huang Chen Fall 2010 1 Outline Asynchronous Counter Operation Synchronous Counter Operation Up/Down Synchronous Counters Design of Synchronous Counters Cascaded Counters

More information

摘 要 互 联 网 的 勃 兴 为 草 根 阶 层 书 写 自 我 和 他 人 提 供 了 契 机, 通 过 网 络 自 由 开 放 的 平 台, 网 络 红 人 风 靡 于 虚 拟 世 界 近 年 来, 或 无 心 插 柳, 或 有 意 噱 头, 或 自 我 表 达, 或 幕 后 操 纵, 网 络

摘 要 互 联 网 的 勃 兴 为 草 根 阶 层 书 写 自 我 和 他 人 提 供 了 契 机, 通 过 网 络 自 由 开 放 的 平 台, 网 络 红 人 风 靡 于 虚 拟 世 界 近 年 来, 或 无 心 插 柳, 或 有 意 噱 头, 或 自 我 表 达, 或 幕 后 操 纵, 网 络 上 海 外 国 语 大 学 硕 士 学 位 论 文 论 文 题 目 从 偶 像 符 号 的 消 解 到 消 费 符 号 的 建 构 网 络 红 人 的 形 象 变 迁 研 究 学 科 专 业 传 播 学 届 别 2013 届 姓 名 孙 清 导 师 王 玲 宁 I 摘 要 互 联 网 的 勃 兴 为 草 根 阶 层 书 写 自 我 和 他 人 提 供 了 契 机, 通 过 网 络 自 由 开 放 的

More information

10384 X0115071 UDC The Research For The Actuality And Development Stratagem Of The China Securities Investment Fund (MBA) 2003 11 2003 12 2003 12 2 0 0 3 11 100 1991, WTO Abstract Abstract The Securities

More information

Microsoft Word - (web)_F.1_Notes_&_Application_Form(Chi)(non-SPCCPS)_16-17.doc

Microsoft Word - (web)_F.1_Notes_&_Application_Form(Chi)(non-SPCCPS)_16-17.doc 聖 保 羅 男 女 中 學 學 年 中 一 入 學 申 請 申 請 須 知 申 請 程 序 : 請 將 下 列 文 件 交 回 本 校 ( 麥 當 勞 道 33 號 ( 請 以 A4 紙 張 雙 面 影 印, 並 用 魚 尾 夾 夾 起 : 填 妥 申 請 表 並 貼 上 近 照 小 學 五 年 級 上 下 學 期 成 績 表 影 印 本 課 外 活 動 表 現 及 服 務 的 證 明 文 件 及

More information

國 立 新 竹 教 育 大 學 音 樂 學 系 音 樂 教 學 碩 士 班 學 位 論 文 新 瓦 屋 客 家 花 鼓 之 研 究 A Research on Hsin-Wa-Wu Hakka Flower-Drum 研 究 生 : 陳 怡 妃 指 導 教 授 : 明 立 國 中 華 民 國 九 十 八 年 三 月 本 論 文 獲 行 政 院 文 化 建 設 委 員 會 文 化 資 產 總 管 理

More information

00. - 0-000 0 10 0 00-0 0 11 12 13 14 15 b 16 17 18 19 0 - 20 0 0-0 0 21 22 H.Mead 0-0 - ( ) 23 ( ) 24 ( ) 25 ( ) 26 27 00 0 00 0 28 29 30 31 ( ) 0 0 32 ( ) 33 ( ) 34 ( ) 35 ( ) 36 ( ) ( ) Northrop F.S.C.

More information

124 第十三期 Conflicts in the Takeover of the Land in Taiwan after the Sino-Japanese War A Case in the Change of the Japanese Names of the Taiwanese Peopl

124 第十三期 Conflicts in the Takeover of the Land in Taiwan after the Sino-Japanese War A Case in the Change of the Japanese Names of the Taiwanese Peopl 123 戰後初期臺灣土地接收的糾紛 以更改日式姓名的臺人遭遇為例 124 第十三期 Conflicts in the Takeover of the Land in Taiwan after the Sino-Japanese War A Case in the Change of the Japanese Names of the Taiwanese People Abstract By Ho Fung-jiao

More information

: : : : : ISBN / C53:H : 19.50

: : : : : ISBN / C53:H : 19.50 : : : : 2002 1 1 2002 1 1 : ISBN 7-224-06364-9 / C53:H059-53 : 19.50 50,,,,,,, ; 50,,,,,,,, 1 ,,,,,,,,,,,,,, ;,,,,,,,,, 2 ,,,, 2002 8 3 ( 1 ) ( 1 ) Deduction One Way of Deriving the Meaning of U nfamiliar

More information

Microsoft PowerPoint - ATF2015.ppt [相容模式]

Microsoft PowerPoint - ATF2015.ppt [相容模式] Improving the Video Totalized Method of Stopwatch Calibration Samuel C.K. Ko, Aaron Y.K. Yan and Henry C.K. Ma The Government of Hong Kong Special Administrative Region (SCL) 31 Oct 2015 1 Contents Introduction

More information

BIBLID 0254-4466(2001)19:1 pp. 249-276 19 1 90 6 ** * ** 88 I 2000 8 249 250 19 1 251 1873-1929 1900 1 1902 1 35 1900 1960 7-12 252 19 1 2 3 2 1900 1902 3 2000 129-197 253 4 5 6 4 1902 1962 103 5 Joseph

More information

Improved Preimage Attacks on AES-like Hash Functions: Applications to Whirlpool and Grøstl

Improved Preimage Attacks on AES-like Hash Functions: Applications to Whirlpool and Grøstl SKLOIS (Pseudo) Preimage Attack on Reduced-Round Grøstl Hash Function and Others Shuang Wu, Dengguo Feng, Wenling Wu, Jian Guo, Le Dong, Jian Zou March 20, 2012 Institute. of Software, Chinese Academy

More information

2008 Nankai Business Review 61

2008 Nankai Business Review 61 150 5 * 71272026 60 2008 Nankai Business Review 61 / 62 Nankai Business Review 63 64 Nankai Business Review 65 66 Nankai Business Review 67 68 Nankai Business Review 69 Mechanism of Luxury Brands Formation

More information

104王三慶.doc

104王三慶.doc 2005 12 95 120 4 2 437-464 On Establishment and Abolishment of 95 New Words made by Wu Tse-t ien, as well as Multiple Version of Written Languages Wang San-Ching Professor, Department of Chinese Literature,

More information

東吳大學

東吳大學 律 律 論 論 療 行 The Study on Medical Practice and Coercion 林 年 律 律 論 論 療 行 The Study on Medical Practice and Coercion 林 年 i 讀 臨 療 留 館 讀 臨 律 六 礪 讀 不 冷 療 臨 年 裡 歷 練 禮 更 老 林 了 更 臨 不 吝 麗 老 劉 老 論 諸 見 了 年 金 歷 了 年

More information

高中英文科教師甄試心得

高中英文科教師甄試心得 高 中 英 文 科 教 師 甄 試 心 得 英 語 學 系 碩 士 班 林 俊 呈 高 雄 市 立 高 雄 高 級 中 學 今 年 第 一 次 參 加 教 師 甄 試, 能 夠 在 尚 未 服 兵 役 前 便 考 上 高 雄 市 立 高 雄 高 級 中 學 專 任 教 師, 自 己 覺 得 很 意 外, 也 很 幸 運 考 上 後 不 久 在 與 雄 中 校 長 的 會 談 中, 校 長 的 一 句

More information

The Development of Color Constancy and Calibration System

The Development of Color Constancy and Calibration System The Development of Color Constancy and Calibration System The Development of Color Constancy and Calibration System LabVIEW CCD BMP ii Abstract The modern technologies develop more and more faster, and

More information

從詩歌的鑒賞談生命價值的建構

從詩歌的鑒賞談生命價值的建構 Viktor E. Frankl (logotherapy) (will-to-meaning) (creative values) Ture (Good) (Beauty) (experiential values) (attitudinal values) 1 2 (logotherapy) (biological) (2) (psychological) (3) (noölogical) (4)

More information

:

: A Study of Huangtao : I Abstract Abstract This text focuses on the special contribution of Huangtao in the history of literature and culture of Fukien, by analyzing the features of Huangtao s thought,

More information

STEAM STEAM STEAM ( ) STEAM STEAM ( ) 1977 [13] [10] STEM STEM 2. [11] [14] ( )STEAM [15] [16] STEAM [12] ( ) STEAM STEAM [17] STEAM STEAM STEA

STEAM STEAM STEAM ( ) STEAM STEAM ( ) 1977 [13] [10] STEM STEM 2. [11] [14] ( )STEAM [15] [16] STEAM [12] ( ) STEAM STEAM [17] STEAM STEAM STEA 2017 8 ( 292 ) DOI:10.13811/j.cnki.eer.2017.08.017 STEAM 1 1 2 3 4 (1. 130117; 2. + 130117; 3. 130022;4. 518100) [ ] 21 STEAM STEAM STEAM STEAM STEAM STEAM [ ] STEAM ; ; [ ] G434 [ ] A [ ] (1970 ) E-mail:ddzhou@nenu.edu.cn

More information

<4D6963726F736F667420576F7264202D205F4230365FB942A5CEA668B443C5E9BB73A740B5D8A4E5B8C9A552B1D0A7F75FA6BFB1A4ACFC2E646F63>

<4D6963726F736F667420576F7264202D205F4230365FB942A5CEA668B443C5E9BB73A740B5D8A4E5B8C9A552B1D0A7F75FA6BFB1A4ACFC2E646F63> 運 用 多 媒 體 製 作 華 文 補 充 教 材 江 惜 美 銘 傳 大 學 應 用 中 文 系 chm248@gmail.com 摘 要 : 本 文 旨 在 探 究 如 何 運 用 多 媒 體, 結 合 文 字 聲 音 圖 畫, 製 作 華 文 補 充 教 材 當 我 們 在 進 行 華 文 教 學 時, 往 往 必 須 透 過 教 案 設 計, 並 製 作 補 充 教 材, 方 能 使 教 學

More information

致 谢 本 论 文 能 得 以 完 成, 首 先 要 感 谢 我 的 导 师 胡 曙 中 教 授 正 是 他 的 悉 心 指 导 和 关 怀 下, 我 才 能 够 最 终 选 定 了 研 究 方 向, 确 定 了 论 文 题 目, 并 逐 步 深 化 了 对 研 究 课 题 的 认 识, 从 而 一

致 谢 本 论 文 能 得 以 完 成, 首 先 要 感 谢 我 的 导 师 胡 曙 中 教 授 正 是 他 的 悉 心 指 导 和 关 怀 下, 我 才 能 够 最 终 选 定 了 研 究 方 向, 确 定 了 论 文 题 目, 并 逐 步 深 化 了 对 研 究 课 题 的 认 识, 从 而 一 中 美 国 际 新 闻 的 叙 事 学 比 较 分 析 以 英 伊 水 兵 事 件 为 例 A Comparative Analysis on Narration of Sino-US International News Case Study:UK-Iran Marine Issue 姓 名 : 李 英 专 业 : 新 闻 学 学 号 : 05390 指 导 老 师 : 胡 曙 中 教 授 上 海

More information

考試學刊第10期-內文.indd

考試學刊第10期-內文.indd misconception 101 Misconceptions and Test-Questions of Earth Science in Senior High School Chun-Ping Weng College Entrance Examination Center Abstract Earth Science is a subject highly related to everyday

More information

A Study on the Relationships of the Co-construction Contract A Study on the Relationships of the Co-Construction Contract ( ) ABSTRACT Co-constructio in the real estate development, holds the quite

More information

01 招 生 简 章 03 考 试 说 明 04 笔 试 样 题 2 emba.pbcsf.tsinghua.edu.cn

01 招 生 简 章 03 考 试 说 明 04 笔 试 样 题 2 emba.pbcsf.tsinghua.edu.cn 01 招 生 简 章 03 考 试 说 明 04 笔 试 样 题 2 emba.pbcsf.tsinghua.edu.cn 清 华 五 道 口 金 融 EMBA 招 生 简 章 金 融 EMBA 教 育 中 心 2012 年, 为 加 快 现 代 金 融 学 科 建 设, 培 养 高 端 金 融 人 才, 促 进 金 融 界 与 金 融 教 育 界 的 联 系, 提 高 金 融 研 究 水 平, 推

More information

Microsoft Word - 24217010311110028谢雯雯.doc

Microsoft Word - 24217010311110028谢雯雯.doc HUAZHONG AGRICULTURAL UNIVERSITY 硕 士 学 位 论 文 MASTER S DEGREE DISSERTATION 80 后 女 硕 士 生 择 偶 现 状 以 武 汉 市 七 所 高 校 为 例 POST-80S FEMALE POSTGRADUATE MATE SELECTION STATUS STUDY TAKE WUHAN SEVEN UNIVERSITIES

More information

1505.indd

1505.indd 上 海 市 孙 中 山 宋 庆 龄 文 物 管 理 委 员 会 上 海 宋 庆 龄 研 究 会 主 办 2015.05 总 第 148 期 图 片 新 闻 2015 年 9 月 22 日, 由 上 海 孙 中 山 故 居 纪 念 馆 台 湾 辅 仁 大 学 和 台 湾 图 书 馆 联 合 举 办 的 世 纪 姻 缘 纪 念 孙 中 山 先 生 逝 世 九 十 周 年 及 其 革 命 历 程 特 展

More information

LH_Series_Rev2014.pdf

LH_Series_Rev2014.pdf REMINDERS Product information in this catalog is as of October 2013. All of the contents specified herein are subject to change without notice due to technical improvements, etc. Therefore, please check

More information

< F5FB77CB6BCBD672028B0B6A46AABE4B751A874A643295F5FB8D5C5AA28A668ADB6292E706466>

< F5FB77CB6BCBD672028B0B6A46AABE4B751A874A643295F5FB8D5C5AA28A668ADB6292E706466> A A A A A i A A A A A A A ii Introduction to the Chinese Editions of Great Ideas Penguin s Great Ideas series began publication in 2004. A somewhat smaller list is published in the USA and a related, even

More information

WTO

WTO 10384 X0115018 UDC MBA 2004 5 14 2004 6 1 WTO 2004 2006 7 2 Abstract According to the promise after our country enter into WTO, our country will open the readymade oil retail market in the end of 2004

More information

A dissertation for Master s degree Metro Indoor Coverage Systems Analysis And Design Author s Name: Sheng Hailiang speciality: Supervisor:Prof.Li Hui,

A dissertation for Master s degree Metro Indoor Coverage Systems Analysis And Design Author s Name: Sheng Hailiang speciality: Supervisor:Prof.Li Hui, 中 国 科 学 技 术 大 学 工 程 硕 士 学 位 论 文 地 铁 内 移 动 通 信 室 内 覆 盖 分 析 及 应 用 作 者 姓 名 : 学 科 专 业 : 盛 海 亮 电 子 与 通 信 导 师 姓 名 : 李 辉 副 教 授 赵 红 媛 高 工 完 成 时 间 : 二 八 年 三 月 十 日 University of Science and Technology of Ch A dissertation

More information

Microsoft Word - 104蔡孟珍.doc

Microsoft Word - 104蔡孟珍.doc 談 成 大 中 文 學 報 第 三 十 二 期 2011 年 3 月 頁 87-124 國 立 成 功 大 學 中 文 系 從 明 清 縮 編 版 到 現 代 演 出 版 牡 丹 亭 崑 劇 重 構 的 幾 個 關 鍵 蔡 孟 珍 * 摘 要 戲 劇 為 搬 演 而 設 格 高 調 雅 的 文 士 劇 本, 宜 於 案 頭 清 玩, 卻 未 必 適 合 場 上 表 演 玉 茗 四 夢, 曲 壇 向

More information

國 史 館 館 刊 第 23 期 Chiang Ching-kuo s Educational Innovation in Southern Jiangxi and Its Effects (1941-1943) Abstract Wen-yuan Chu * Chiang Ching-kuo wa

國 史 館 館 刊 第 23 期 Chiang Ching-kuo s Educational Innovation in Southern Jiangxi and Its Effects (1941-1943) Abstract Wen-yuan Chu * Chiang Ching-kuo wa 國 史 館 館 刊 第 二 十 三 期 (2010 年 3 月 ) 119-164 國 史 館 1941-1943 朱 文 原 摘 要 1 關 鍵 詞 : 蔣 經 國 贛 南 學 校 教 育 社 會 教 育 掃 盲 運 動 -119- 國 史 館 館 刊 第 23 期 Chiang Ching-kuo s Educational Innovation in Southern Jiangxi and

More information

Microsoft Word - 武術合併

Microsoft Word - 武術合併 11/13 醫 學 系 一 年 級 張 雲 筑 武 術 課 開 始, 老 師 並 不 急 著 帶 我 們 舞 弄 起 來, 而 是 解 說 著 支 配 氣 的 流 動 為 何 構 成 中 國 武 術 的 追 求 目 標 武 術, 名 之 為 武 恐 怕 與 其 原 本 的 精 義 有 所 偏 差 其 實 武 術 是 為 了 讓 學 習 者 能 夠 掌 握 身 體, 保 養 身 體 而 發 展, 並

More information

IP TCP/IP PC OS µclinux MPEG4 Blackfin DSP MPEG4 IP UDP Winsock I/O DirectShow Filter DirectShow MPEG4 µclinux TCP/IP IP COM, DirectShow I

IP TCP/IP PC OS µclinux MPEG4 Blackfin DSP MPEG4 IP UDP Winsock I/O DirectShow Filter DirectShow MPEG4 µclinux TCP/IP IP COM, DirectShow I 2004 5 IP TCP/IP PC OS µclinux MPEG4 Blackfin DSP MPEG4 IP UDP Winsock I/O DirectShow Filter DirectShow MPEG4 µclinux TCP/IP IP COM, DirectShow I Abstract The techniques of digital video processing, transferring

More information

2012 2 157 No. 2 2012 Jinan Journal Philosophy and Social Sciences Sum No. 157 100871 I127. 41 A 1000-5072 2012 02-0002 - 10 4 1 1 2 1540 1503 2 2011-06 - 27 1969 1 1985 19 2 3 1997 1998 3 2 34 2 3 1540

More information

廣州舊城區的保護和發展

廣州舊城區的保護和發展 黄 爲 爲 爲 爲 塲 爲 茘 爲 爲 爲 爲 産 统 装 爲 爲 爲 爲 爲 爲 爲 茘 衞 産 爲 塲 爲 爲 爲 爲 爲 滙 爲 滙 爲 爲 爲 爲 爲麽麽 强 迹 啓 啓 産 産 [1] [2] [3] [4] 20 [5] 2000 [6] [7], 20 [8] http://www.gz.gov.cn2002.7.14 [9] 2000.4 [10] [11] 20 2001.1.29

More information

6張謇教育思想之研究.doc

6張謇教育思想之研究.doc 1. (74)16 5-30 2. (61) 3. (84) : 4. (51) 5. (76) 6. (52) 7. (82) - : 267-276 :: 8. (86) 9. (84) 140 243-255 10. (88) 11. (86) 115-144 12. (51) 408-425 13. (6) 33-36 14. (67) 15. (19) 16. ( ) (72) 17. (75)

More information

入學考試網上報名指南

入學考試網上報名指南 入 學 考 試 網 上 報 名 指 南 On-line Application Guide for Admission Examination 16/01/2015 University of Macau Table of Contents Table of Contents... 1 A. 新 申 請 網 上 登 記 帳 戶 /Register for New Account... 2 B. 填

More information

國立中山大學學位論文典藏

國立中山大學學位論文典藏 i Examinations have long been adopting for the selection of the public officials and become an essential tradition in our country. For centuries, the examination system, incorporated with fairness, has

More information

Microsoft Word - 11月電子報1130.doc

Microsoft Word - 11月電子報1130.doc 發 行 人 : 楊 進 成 出 刊 日 期 2008 年 12 月 1 日, 第 38 期 第 1 頁 / 共 16 頁 封 面 圖 話 來 來 來, 來 葳 格 ; 玩 玩 玩, 玩 數 學 在 11 月 17 到 21 日 這 5 天 裡 每 天 一 個 題 目, 孩 子 們 依 據 不 同 年 段, 尋 找 屬 於 自 己 的 解 答, 這 些 數 學 題 目 和 校 園 情 境 緊 緊 結

More information

untitled

untitled Co-integration and VECM Yi-Nung Yang CYCU, Taiwan May, 2012 不 列 1 Learning objectives Integrated variables Co-integration Vector Error correction model (VECM) Engle-Granger 2-step co-integration test Johansen

More information

(65 ),,, :,,, 2500, 25 :4161 ; 180,, 3782, 54,,,,, : (b i, b i) (l ng,l ng) (b ig n,b ig nr) (ji gu,ji gu ) (m nd o,m ndao) (hu,hu ) (ch is n,c

(65 ),,, :,,, 2500, 25 :4161 ; 180,, 3782, 54,,,,, : (b i, b i) (l ng,l ng) (b ig n,b ig nr) (ji gu,ji gu ) (m nd o,m ndao) (hu,hu ) (ch is n,c : ( ) ( ) Ξ ( ) :1) ;2) ; ; ; 1992 6, 10,,,,,,, ( ) ( ),?? ( ), (,1992),,, ( ) ( ) 7 ( ), Ξ 2002 2003 (:02JAZJD740006) (:HBK01-05/ 015), 73 2003 3 (65 ),,, :,,, 2500, 25 :4161 ; 180,, 3782, 54,,,,, : (b

More information

PowerPoint 簡報

PowerPoint 簡報 國 立 中 山 大 學 財 務 管 理 學 系 Department of Finance, NSYSU 刊 登 之 文 章, 其 版 權 屬 於 本 系, 非 徵 得 同 意, 不 得 轉 載 本 期 目 錄 1. 財 管 系 榮 譽 及 活 動 訊 息 2. 讓 夢 想 貣 飛 / 楊 佳 霖 3. Student Ambassador(SA)/ 陳 伯 穎 4. 財 務 報 表 之 重 大 變

More information

BC04 Module_antenna__ doc

BC04 Module_antenna__ doc http://www.infobluetooth.com TEL:+86-23-68798999 Fax: +86-23-68889515 Page 1 of 10 http://www.infobluetooth.com TEL:+86-23-68798999 Fax: +86-23-68889515 Page 2 of 10 http://www.infobluetooth.com TEL:+86-23-68798999

More information

PowerPoint Presentation

PowerPoint Presentation Decision analysis 量化決策分析方法專論 2011/5/26 1 Problem formulation- states of nature In the decision analysis, decision alternatives are referred to as chance events. The possible outcomes for a chance event

More information

THE APPLICATION OF ISOTOPE RATIO ANALYSIS BY INDUCTIVELY COUPLED PLASMA MASS SPECTROMETER A Dissertation Presented By Chaoyong YANG Supervisor: Prof.D

THE APPLICATION OF ISOTOPE RATIO ANALYSIS BY INDUCTIVELY COUPLED PLASMA MASS SPECTROMETER A Dissertation Presented By Chaoyong YANG Supervisor: Prof.D 10384 070302 9825042 UDC 2001.6. 2001.7. 20016 THE APPLICATION OF ISOTOPE RATIO ANALYSIS BY INDUCTIVELY COUPLED PLASMA MASS SPECTROMETER A Dissertation Presented By Chaoyong YANG Supervisor: Prof.Dr. Xiaoru

More information

5 1 linear 5 circular 2 6 2003 10 3000 2 400 ~ 500 4500 7 2013 3500 400 ~ 500 8 3 1900 1. 2 9 1 65

5 1 linear 5 circular 2 6 2003 10 3000 2 400 ~ 500 4500 7 2013 3500 400 ~ 500 8 3 1900 1. 2 9 1 65 2014 2 43 EXAMINATIONS RESEARCH No. 2 2014 General No. 43 李 迅 辉 随 着 新 课 程 改 革 的 不 断 深 入, 教 学 理 念 逐 步 更 新, 学 生 的 英 语 水 平 也 在 逐 渐 提 高, 但 沿 用 多 年 的 高 考 英 语 书 面 表 达 的 评 分 标 准 并 没 有 与 时 俱 进, 已 经 不 能 完 全 适 应

More information

2005 4,,,,,,,,,,,,, (2001 ) ;() ( 1997 ) ; ( 1997 ) ; () (1996 ) ;,: ( 1995 ) ; (1995 ),,,, (J13) (J9) (J10), (J19) (J17) 58

2005 4,,,,,,,,,,,,, (2001 ) ;() ( 1997 ) ; ( 1997 ) ; () (1996 ) ;,: ( 1995 ) ; (1995 ),,,, (J13) (J9) (J10), (J19) (J17) 58 ,,,, 44 %,,,,,,,,,,,, 57 2005 4,,,,,,,,,,,,, (2001 ) ;() ( 1997 ) ; ( 1997 ) ; () (1996 ) ;,: ( 1995 ) ; (1995 ),,,, (J13) (J9) (J10), (J19) (J17) 58 ,,,,,,,,,,, 1945 11 5,, ( ) : ; 1945 11,,,,,:, 1987,802,

More information

Windows XP

Windows XP Windows XP What is Windows XP Windows is an Operating System An Operating System is the program that controls the hardware of your computer, and gives you an interface that allows you and other programs

More information

声 明 本 人 郑 重 声 明 : 此 处 所 提 交 的 硕 士 学 位 论 文 基 于 等 级 工 鉴 定 的 远 程 考 试 系 统 客 户 端 开 发 与 实 现, 是 本 人 在 中 国 科 学 技 术 大 学 攻 读 硕 士 学 位 期 间, 在 导 师 指 导 下 进 行 的 研 究

声 明 本 人 郑 重 声 明 : 此 处 所 提 交 的 硕 士 学 位 论 文 基 于 等 级 工 鉴 定 的 远 程 考 试 系 统 客 户 端 开 发 与 实 现, 是 本 人 在 中 国 科 学 技 术 大 学 攻 读 硕 士 学 位 期 间, 在 导 师 指 导 下 进 行 的 研 究 中 国 科 学 技 术 大 学 硕 士 学 位 论 文 题 目 : 农 村 电 工 岗 位 培 训 考 核 与 鉴 定 ( 理 论 部 分 ) 的 计 算 机 远 程 考 试 系 统 ( 服 务 器 端 ) 的 开 发 与 实 现 英 文 题 目 :The Realization of Authenticating Examination System With Computer & Web for

More information

Microsoft Word - D-2°w¶Ë¬ì¹ï¤U�Iµh®{¤âÀˬd¬yµ{_¬x°ö�×__P.329-335_.doc

Microsoft Word - D-2°w¶Ë¬ì¹ï¤U�Iµh®{¤âÀˬd¬yµ{_¬x°ö�×__P.329-335_.doc 針 傷 科 對 下 背 痛 徒 手 檢 查 流 程 329 針 傷 科 對 下 背 痛 徒 手 檢 查 流 程 洪 培 修 1 張 晉 賢 2 張 世 良 3 1 嘉 義 基 督 教 醫 院 中 醫 科 系 2 長 庚 紀 念 醫 院 中 醫 醫 院 基 隆 分 院 暨 長 庚 大 學 醫 學 院 3 中 國 醫 藥 大 學 中 醫 院 學 針 灸 研 究 所 摘 要 前 言 腰 痛 或 下 背 痛

More information

untitled

untitled and Due Diligence M&A in China Prelude and Due Diligence A Case For Proper A Gentleman s Agreement? 1 Respect for the Rule of Law in China mandatory under law? CRITICAL DOCUMENTS is driven by deal structure:

More information

01何寄澎.doc

01何寄澎.doc 1 * ** * ** 2003 11 1-36 Imitation and the Formation and Interpretation of the Canon: Lu Chi s Nigushi Ho Chi-p eng Professor, Department of Chinese Literature, National Taiwan University. Hsu Ming-ch

More information

Microsoft PowerPoint - IAS 21 - IFRS宣導會.pptx

Microsoft PowerPoint - IAS 21 - IFRS宣導會.pptx IAS 21 Nov 19, 2010 Agenda Page 1 1 2 4 3 11 4 17 5 IFRS 23 Section 1 Section 1 WHY IAS 21? IAS 21 2 Section 1 Determination Functional Currency Presentation Currency First Time Adoption IFRS IAS 21 2

More information

<4D6963726F736F667420576F7264202D2035B171AB73B6CBA8ECAB73A6D3A4A3B6CBA158B3AFA46CA9F9BB50B169A445C4D6AABAB750B94AB8D6B9EFA4F1ACE3A873>

<4D6963726F736F667420576F7264202D2035B171AB73B6CBA8ECAB73A6D3A4A3B6CBA158B3AFA46CA9F9BB50B169A445C4D6AABAB750B94AB8D6B9EFA4F1ACE3A873> 中 正 漢 學 研 究 2012 年 第 一 期 ( 總 第 十 九 期 ) 2012 年 6 月 頁 111~134 國 立 中 正 大 學 中 國 文 學 系 111 從 哀 傷 到 哀 而 不 傷 : 陳 子 昂 與 張 九 齡 的 感 遇 詩 對 比 研 究 * 丁 涵 摘 要 在 中 國 古 典 文 學 語 境 中, 一 個 主 題 的 奠 立 往 往 需 要 歷 時 彌 久, 而 這 本

More information

從篤加有二「區」談當代平埔文化復振現相

從篤加有二「區」談當代平埔文化復振現相 從 篤 加 有 二 邱 談 族 群 正 名 運 動 從 篤 加 有 二 邱 談 族 群 正 名 運 動 陳 榮 輝 台 南 女 子 技 術 學 院 通 識 教 育 中 心 講 師 摘 要 本 文 從 篤 加 村 非 平 埔 族 裔 的 正 名 運 動, 探 討 篤 加 村 民 因 不 認 同 廟 後 區 ( 邱 ) 所 形 成 的 平 埔 族 裔 概 念, 從 地 理 變 遷 村 廟 沿 革 族 譜

More information