<<

Induction of Root and Pattern Lexicon for Unsupervised Morphological Analysis of

Bilal Khaliq John Carroll Dept of Informatics, University of Sussex Dept of Informatics, University of Sussex Brighton BN1 9QJ, UK Brighton BN1 9QJ, UK [email protected] [email protected]

Abstract In this paper we describe a conceptually simple yet effective unsupervised approach to We propose an unsupervised approach to learning non-concatenative . We learning non-concatenative morphology, apply our approach to inducing an Arabic which we apply to induce a lexicon of Arabic roots and pattern templates. The lexicon of trilateral roots and pattern templates. approach is based on the idea that roots and Lexicon acquisition is based on the idea that patterns may be revealed through mutually roots and patterns may be revealed by their recursive scoring based on hypothesized converses, i.e. roots are identified from pattern and root frequencies. After a further occurrences of patterns and conversely patterns iterative refinement stage, morphological are recognized from root frequencies. analysis with the induced lexicon achieves a Subsequently, the lexicons are iteratively root identification accuracy of over 94%. improved by refining the strengths Our approach differs from previous work on computed in the previous step. unsupervised learning of Arabic The paper is organized as follows. We survey morphology in that it is applicable to naturally-written, unvowelled text. previous related work (Section 2), and then give a brief introduction to Arabic root and pattern 1 Introduction morphology (Section 3). We explain our basic technique for unsupervised lexicon induction in Manual development of morphological analysis Section 4, followed by the refinement procedure systems is expensive. It is impractical to develop (Section 5). Section 6 describes how the lexicon morphological descriptions for more than a very is used for morphological analysis. Finally, we small proportion of human . In recent present an evaluation (Section 7) and conclusions years a number of approaches have been (Section 8). proposed that learn the morphology of a from unannotated text. The Morpho 2 Related Work Challenge and similar competitions have further motivated researchers to devise techniques for Beesley (1996) describes one of the first unsupervised learning of language. morphological analysis systems for Arabic, Previous work in unsupervised morphology based on finite-state techniques with manually learning has mostly addressed concatenative acquired lexicons and rules. This kind of morphology, in which surface forms are approach, although potentially producing an sequentially separated or segmented into efficient and accurate system, is expensive in morpheme units. However, some languages (in time and linguistic expertise, and lacks particular ) have another type robustness in terms of extendibility to word types of word formation in which combine not in the (Ahmed, 2000). in a non-concatenative manner, through the Darwish (2002) describes a semi-automatic interdigitation of a root morpheme with an affix technique that learns morphemes and induces or pattern template. Unsupervised learning of rules for deriving stems using an existing non-concatenative morphology has received dictionary of word and root pairs. It is an easy to comparatively little attention. build and fairly robust method of performing morphological analysis. Clark (2007) investigates semi-supervised learning using the

1012 International Joint Conference on Natural Language Processing, pages 1012–1016, Nagoya, Japan, 14-18 October 2013. complex broken plural structure of Arabic as a where ranges from 1 to . For example, the test case. He employs memory-based algorithms, decomposition of the word yErf is shown in ݇ .with the aim of gaining insights into human Figureݔ 1 . , Other researchers have applied statistical and , 〈݂− − − ,ݎܧinformation-theoretic approaches to 〈ݕ ⎧ , ⎫ ⎪ ,〈−ݎ − − ,,݂ܧunsupervised learning of morphology from raw ⎪ 〈ݕ 〈− − ܧ− ,݂ݎݕ〉 → ݂ݎܧunannotated) text corpora. Goldsmith (2000, ݕ) 2006) and Cruetz and Lagus (2005, 2007) use the ⎨ ⎬ ⎪ 〈− − − ݕ ݂ݎܧ〉 ⎪ of− a− word− − into〉 ⎭ all ,݂ݎܧMinimum Description Length (MDL) principle, Figure 1. Decomposition⎩〈ݕ considering input data to be ‘compressed’ into a possible combinations of roots and patterns. morphologically analysed representation. An alternative perspective adopted by Schone and Jurafsky (2001) induces semantic relatedness 4 Building Lexicons Using Contrastive between word pairs by Latent Semantic Indexing. Scoring Most work on unsupervised learning of morphology has focused on concatenative Based on the idea that roots and patterns may be morphology (Hammarström and Borin, 2011). revealed by their converses, we score a pattern Studies that have focussed on non-concatenative based on the frequency of occurrence of the roots, morphology include that of Rodriguez and Ćavar and score a root according to the number of (2005), who learn roots from artificially occurrences of patterns. We score each generated text using a number of orthographic morpheme and then rescore it weighted by heuristics, and then apply constraint-based previous scores. Our technique resembles the learning to improve the quality of the roots. hubs and authorities algorithm originally devised Xanthos (2008) deciphers roots and patterns for rating Web pages (Kleinberg, 1999), which from phonetic transcriptions of Arabic text, using assigns to each Web page two scores: its hub MDL to refine the root and pattern structures. value and its authority. These two values are Our work differs from these previous updated in a similar mutually recursive manner approaches in that (1) we learn intercalated as we describe for roots and patterns. morphology, identifying the root and transfixes/ incomplete pattern for , and (2) we start 4.1 Frequency-Based Scoring from ‘natural’ text without short or The initial scoring function is simple: firstly, we diacritical markers. aggregate over the number of occurrences of a root radical sequence in a word w , for words 3 Root and Pattern Morphology i i=1,2,…N in the input dataset. The function for Words in Arabic are formed through three scoring each pattern in the target word, t, is given morphological processes: (i) fusion of a root in equation (Eq. 2). form and pattern template to derive a base word; (ii) affixation, including inflectional morphemes ( ) = ே 1 = marking gender, plurality and/or tense, resulting ௫ ௫ ௬ ೔ in a stem; and (iii) possible attachment of a final ܵ ݌௧ ෍ ൫ หݎ௧ ݎ௪ ൯ (Eq. 2) ௜ୀଵ layer of clitics. Our work addresses the first two The function for scoring the root, , in each of these processes. target word, t, with pattern, , is given in As an example of word formation in Arabic, ௫ equation (Eq. 3). ݎ௧ the word ktAby is formed from the root Ktb and ௫ ݌௧ the pattern --A-y, where y is an inflectional marker and A is the derivational infix marker for ( ) = 1 = nouns. ே ௫ ௫ ௬ During analysis, we decompose each word ೔ ܵ ݎ௧ ෍ ൫ ห݌௧ ݌௪ ൯ (Eq. 3) into a set of tuples encoding all possible ௜ୀଵ combinations of a root (of at least 3 letters) andݓ We choose this as our baseline, to which we associated pattern (Eq. 1) ݇ compare subsequent enhancements. ( { , } ௫ ௫ (Eq. 1) 〈 ݌ ݎ〉 → (ݓ ݀ 1013 4.2 Scaling obtained from the frequency-based method (Section 4) are frequency counts or weighted Since pattern strength is computed based on root frequency counts. The scoring and rescoring in occurrence frequency and vice-versa, each this refinement step differs in that it evaluates pattern and root has a different score range due each root by averaging over scores of the to the distinct distributions of patterns and roots. corresponding patterns it occurs with in the In order to make the scores comparable and dataset. We again iteratively refine based on the contribute equally, we scale the scores in one previous scores for k=0,1,2,…m iterations, lexicon with respect to the other. We take the pattern lexicon as reference and 1 scale each root, ru (u=1,2,…R entries in root ( ) = = lexicon) by the ratio of the maximum pattern ே ௫ ௫ (score to the maximum root score: ௝ ௧ ௞ିଵ ݓ݅ ௧ ݓ(Eq.݅ 7 ܵ ݎ ௥ ෍ ൫ܵ ൫݌ ൯หݎ ݎ ൯ ݂ ௜ୀଵ max( ( )) where is the number of words with root r, from ( ) = ( )(× max( ( )) a total of N vocabulary words. Similarly, for the ௥ ܵ ݌ (Eq. 4) pattern݂ rescoring with the best so far pattern, , ܵܵ ݎ௨ ܵ ݎ௨ ܵ ݎ ௕ 1 ௪ 4.3 Iterative Rescoring ( ) = ே = ݌ ௫ ௫ Having obtained initial scores for the root and ௝ ௪ ௞ିଵ ݓ݅ ௪ ݓ݅ ܵ ݌ ௣ ෍ ൫ܵ ൫ݎ ൯ห݌ ݌ ൯(Eq. 8) pattern lexicons, they are improved through an ݂ ௜ୀଵ iterative rescoring process. We rescore each Here we sum over the score of counterpart morpheme lexicon in a similar manner to morphemes based on the match of the target equations (Eq. 2) and (Eq. 3), but weighted with morpheme in a vocabulary word, unlike the the normalized score for each morpheme of rescoring step, where we match the previous scores. This is an iterative process corresponding roots. starting with the initial score, S0, calculated using frequency counts (as in section 4.1). Let Sj be the 6 Morphological Analysis new score based on previous scores, and scaled A word, w , is analysed into its potential root and scores, Sj-1 and SSj-1, respectively, for iterations i j=0,1,2,…n, pattern template by considering every possible combination of trilateral root and corresponding pattern pairs, , , as defined in equation (Eq. ( ) = ே ( ) max( ) = 1). Each analysis௫ is௫ scored with the sum of the ௫ ௫ ௫ ݕ ௝ ௝ିଵ ௝ିଵ scores for the〈ݎ root,݌ 〉 , and pattern, , in the (௧ ෍ ቀܵ ݌௧ ⁄ ܵ ቚ݌௧ ݌ݓ(Eq.݅ቁ 5ݎ ܵ ௜ୀଵ root lexicon and pattern௫ lexicon, respectively.௫ While combining scoresݎ we again apply݌ scaling

( ) = ே ( ) max( ) = as in equation (Eq. 4) in order to guarantee equal ,௫ ௫ ௫ ݕ contributions from each morpheme. The analysis (௧ (Eq.ݓ݅൯ 6ݎ௧ ⁄ ܵܵ௝ିଵ หݎ ௝ ݌௧ ෍ ൫ܵܵ௝ିଵܵ ௜ୀଵ ݎ x, with the highest score, as calculated in equation (Eq. 9), is selected and is output. Here we have normalized the score with respect to the maximum value for the reference pattern max ( ( ) + ( ) ) lexicon, thus keeping the magnitude of the .. ௫ ௫ rescored value in range. ௪  ௪ (Eq. 9) ௫ୀଵ ௡ ܵ ݎ ܵܵ ݌ 5 Refinement Since we are considering text without diacritics, due to the absence of short vowels, we The refinement phase considers the overall only expect words to contain single letter infixes. strength of occurrence of each morpheme in Hence we also experiment with an alternative the vocabulary. Thus, if a certain root morpheme configuration of the word decomposition, is a true morpheme then all the pattern , in which only those tuples with single morphemes it occurs with would have higher character௭ ௭ infixes in patterns are considered for scores since they also would be true morphemes. analysis,〈ݎ ݌ 〉 and all other tuples are dropped. We In such a case, this phase would increase the refer to this configuration as ‘IF1’ in the overall average strength for the root. The scores following evaluation.

1014 7 Evaluation the rescoring (RSc) and refinement (Ref) phases, with n=5 and m=5 iterations respectively The evaluation dataset comes from the Quranic 1 (NoP n+m). There is a sudden improvement in Arabic Corpus (QAC), which contains approx- accuracy after the first rescoring phase, and imately 7370 undiacritized, stemmed token types. gradual improvement thereafter until the fifth. Although for evaluation purposes we use the The refinement phase shows a similar trend, with stemmed vocabulary provided by QAC, such a sudden improvement in accuracy at NoP 5+1. stemmed words could be obtained using existing Here too the improvement is more gradual after techniques for unsupervised concatenative each further iteration. morphology learning (e.g. Poon et al., 2009). More than 7192 words (95% of the total Configuration Total Percentage vocabulary) are tagged with their root forms Correct Correct (%) since the consists mostly of words of Baseline (BL) 4055 74.2 derivable forms, with very few proper nouns. RSc_NoP1 4444 81.2 Sometimes alterations in root radicals take place, RSc_NoP5 4539 83.0 for example, in hollow roots, when moving from RSc_Ref_NoP5+1 4940 90.3 a root containing a long to the surface RSc_Ref_NoP5+5 5123 93.6 word, the long vowel might change its form to RSc_Ref_IF1 5159 another type or get dropped. Such words with 94.3 hollow roots or reduplicated radicals, whose Table 1. Results at key stages. characters do not match every radical of the root, were excluded from the evaluation as they are Table 1 shows the number of correct results at beyond the scope of the learning algorithm. After key stages. Rescoring and refinement each these exclusions, 5468 word and root evaluation improve accuracy by 7 percentage points on their pairs remain. first iteration. This shows the advantage of using 7.1 Root Identification weighted morpheme scores. The subsequent iterations give total improvements of We evaluate morphological analysis through approximately 3 points. The IF1 configuration correct identification of the root. Accuracy is yields a further improvement of 0.75 points, measured in terms of the percentage of roots that indicating that some irrelevant analyses have are correctly identified. been filtered out. With all the enhancements, the overall accuracy of 94.3% is an improvement of RSc/Ref more than 20 percentage points over the baseline. 5200 8 Conclusions and Future Directions s i

s 5000 y l a

n 4800 We have presented a novel, unsupervised A

t

c 4600 approach to learning non-concatenative e r r 4400 morphology. The approach learns trilateral roots o C

f 4200 and pattern templates, based on the idea that o

.

o 4000 each may be revealed by their converses, using a N 3800 mutually recursive scoring method. A 1 2 3 4 5 5+1 5+2 5+3 5+4 5+5 subsequent refinement phase further increases RSc/Ref 4055 4444 4510 4532 4539 4940 5028 5094 5101 5123 accuracy. % Acc 74.1 81.2 82.4 82.8 83.0 90.3 91.9 93.1 93.2 93.6 The approach could be extended to roots No. of Iterations (NoP) beyond trilateral by adapting the scoring function to accommodate for morpheme length. In the future, we intend to apply the method to learning Figure 2. Results for iterative scoring and other kinds of morphological structures. refinement. Acknowledgments Using the initial frequency-based scoring We thank the referees for valuable comments, function (Section 4.1), we obtain a baseline (BL) and in particular pointing out the correspondence accuracy of 74.1%. Figure 2 shows the results of to the hubs and authorities algorithm.

1 http://corpus.quran.com/

1015 References Paul Rodrigues and Damir Ćavar. 2005. Learning Arabic morphology using information theory. In Mohamed Attia Ahmed, 2000. A Large-Scale Proceedings of the Chicago Society. Computational Processor of the Arabic Vol 41. Chicago: University of Chicago. 49-58. Morphology, and Applications. Master’s Thesis, Faculty of Engineering, Cairo University, Cairo, Patrick Schone and Daniel Jurafsky. 2001. Egypt. Knowledge-free induction of inflectional morphologies. In Proceedings of the Conference of Kenneth Beesley. 1996. Arabic finite-state the North American Chapter of the ACL, Pittsburgh, morphological analysis and generation. In PA, 183-191. Proceedings of the 16th International Conference on Computational Linguistics (COLING), Aris Xanthos. 2008. Apprentissage Automatique de la Copenhagen, Denmark, 89-94. Morphologie: Le Cas des Structures Racine- Schème ‘The Automatic Learning of Morphology: Alexander Clark. 2007. Supervised and unsupervised The Case of Root-and-Pattern Structures’. Berne, learning of Arabic morphology. Arabic Switzerland: Peter Lang. Computational Morphology, volume 38 of Text, Speech and Language Technology, pages 181-200. Mathias Creutz and Krista Lagus. 2005. Inducing the morphological lexicon of a natural language from unannotated text. In Proceedings of the International and Interdisciplinary Conference on Adaptive Knowledge Representation and Reasoning (AKRR ‘05), 106-113. Mathias Creutz and Krista Lagus. 2007. Unsupervised models for morpheme segmentation and morphology learning. ACM Transactions on Speech and Language Processing, 4(1-3):1-33. Kareem Darwish. 2002. Building a shallow morphological analyzer in one day. In Proceedings of the ACL-02 Workshop on Computational Approaches to Semitic Languages, Philadelphia, PA, 1-8. Guy De Pauw and Peter Wagacha. 2007. Bootstrapping morphological analysis of Gikuyu using unsupervised maximum entropy learning. In Proceedings of the Eighth Annual Conference of the International Speech Communication Association. Antwerp, Belgium. John Goldsmith. 2000. Linguistica: An automatic morphological analyser. In Proceedings of the 36th Meeting of the Chicago Linguistic Society. 125-139. John Goldsmith. 2006. An algorithm for the unsupervised learning of morphology. Natural Language Engineering, 12(4):353-371. Harold Hammarström and Lars Borin. 2011. Unsupervised learning of morphology. Computational Linguistics, 37(2):309-350. Jon Kleinberg. 1999. Authoritative sources in a hyperlinked environment. Journal of the ACM, 46(5):604-632. Hoifung Poon, Colin Cherry and Kristina Toutanova. 2009. Unsupervised morphological segmentation with log-linear models. In Proceedings of the Conference of the North American Chapter of the ACL, Boulder, CO, 209-217.

1016