A Unigram Orientation Model for Statistical Machine Translation

A Unigram Orientation Model for Statistical Machine Translation

A Unigram Orientation Model for Statistical Machine Translation Christoph Tillmann IBM T.J. Watson Research Center Yorktown Heights, NY 10598 [email protected] Abstract b5 airspace In this paper, we present a unigram segmen- b4 tation model for statistical machine transla- Lebanese b tion where the segmentation units are blocks: 3 pairs of phrases without internal structure. The violate b2 segmentation model uses a novel orientation warplanes component to handle swapping of neighbor b1 blocks. During training, we collect block un- Israeli igram counts with orientation: we count how A A A t A A A often a block occurs to the left or to the right of l l l n l l l some predecessor block. The orientation model T H A t m j l is shown to improve translation performance A r s h j w b } over two models: 1) no block re-ordering is b r k A y n r y A l A used, and 2) the block swapping is controlled A P } n only by a language model. We show exper- t y y imental results on a standard Arabic-English l y translation task. P 1 Introduction Figure 1: An Arabic-English block translation example taken from the devtest set. The Arabic words are roman- In recent years, phrase-based systems for statistical ma- ized. chine translation (Och et al., 1999; Koehn et al., 2003; Venugopal et al., 2003) have delivered state-of-the-art performance on standard translation tasks. In this pa- section 3, block swapping where only a trigram language per, we present a phrase-based unigram system similar model is used to compute probabilities between neighbor to the one in (Tillmann and Xia, 2003), which is ex- blocks fails to improve translation performance. (Wu, tended by an unigram orientation model. The units of 1996; Zens and Ney, 2003) present re-ordering models translation are blocks, pairs of phrases without internal that make use of a straight/inverted orientation model that structure. Fig. 1 shows an example block translation us- is related to our work. Here, we investigate in detail ing five Arabic-English blocks ¢¡¢£¥¤¥¤¥¤¥£¦ ¨§ . The unigram the effect of restricting the word re-ordering to neighbor orientation model is trained from word-aligned training block swapping only. data. During decoding, we view translation as a block In this paper, we assume a block generation process that segmentation process, where the input sentence is seg- generates block sequences from bottom to top, one block mented from left to right and the target sentence is gener- at a time. The score of a successor block depends on its ated from bottom to top, one block at a time. A monotone predecessor block © and on its orientation relative to the ¢¡ block sequence is generated except for the possibility to block ¨© . In Fig. 1 for example, block is the predeces- swap a pair of neighbor blocks. The novel orientation sor of block , and block is the predecessor of block ¥© model is used to assist the block swapping: as shown in ¨ . The target clump of a predecessor block is adja- ¢¡ cent to the target clump of a successor block . A right occurs to the left of block . Although the joint block ¨ £¦ ¥¡ © ¢¡ ¨ ¨ adjacent predecessor block is a block where addition- ¡ consisting of the two smaller blocks and ally the source clumps are adjacent and the source clump has not been seen in the training data, we can still profit of ¨© occurs to the right of the source clump of . A left from the fact that block occurs more frequently with adjacent predecessor block is defined accordingly. left than with right orientation. In our Arabic-English ¤£ £¦¥§£ ,) ¨ ¡ ¡©¨ ¨ !4365 ¢¡ During decoding, we compute the score of a training data, block has been seen ¡ times £ ¥ £ ,+ ¨ ¡ ¡ ¨ !87 block sequence with orientation as a product of with left orientation, and ¡ with right orien- block bigram scores: tation, i.e. it is always involved in swapping. This intu- £ ition is formalized using unigram counts with orientation. £ £ £¦¥ ¡ £¦¥ ¡ £ £¦¥ ¨ ¡ ¡ ¨ The orientation model is related to the distortion model ¡ ¢¡ (1) ¡ in (Brown et al., 1993), but we do not compute a block ¥ £¦ £¦ alignment during training. We rather enumerate all rele- where is a block and is a three-valued vant blocks in some order. Enumeration does not allow orientation component linked to the block (the orienta- ¡ ¥ us to capture position dependent distortion probabilities, tion of the predecessor block is ignored.). A block ¥ but we can compute statistics about adjacent block prede- "! has right orientation ( ) if it has a left adjacent cessors. predecessor. Accordingly, a block has left orientation ¥ Our baseline model is the unigram monotone model de- ( #!$ ) if it has a right adjacent predecessor. If a block scribed in (Tillmann and Xia, 2003). Here, we select has neither a left or right adjacent predecessor, its orien- ¥ blocks from word-aligned training data and unigram %! tation is neutral ( ). The neutral orientation is not ¨ block occurrence counts ¡ are computed: all blocks modeled explicitly in this paper, rather it is handled as a for a training sentence pair are enumerated in some order default case as explained§ below. In Fig. 1, the orienta- ¥ £ £¦ £¦ £ ¡ ¨ and we count how often a given block occurs in the par- ! tion sequence is ¡ , i.e. block and 1 ¨§ allel training data . The training algorithm yields a list block are generated using left orientation. During de- 3 ¥ 9 ¨ of about blocks per training sentence pair. In this pa- ! coding most blocks have right orientation ¡ , since per, we make extended use of the baseline enumeration the block translations are mostly monotone. ¤£ £¦¥§£ ¡ ¨ ¡ procedure: for each block , we additionally enumerate © We try to find a block sequence with orientation ¡ £ £¦¥§£ ¡ ¨ ¡ all its left and right predecessors . No optimal block that maximizes ¢¡ . The following three types segmentation is needed to compute the predecessors: for © of parameters are used to model the block bigram score £¦¥ ¡ £¦¥ ¡ ¨ each block , we check for adjacent predecessor blocks ¡ in Eq. 1: that also occur in the enumeration list. We compute left ,) ¨ ¨ & ¡ Two unigram count-based models: ¡ and orientation counts as follows: ¥ ¨ ¨ ¡ ¡ . We compute the unigram probability of (' *) < ¨ ¨ ! : ¡ ¡ a block based on its occurrence count . The 2 © blocks are counted from word-aligned training data. ; right adjacent predecessor of We also collect unigram counts with orientation: a *) ,+ ¨ ¨ ¥© ¡ left count ¡ and a right count . These Here, we enumerate all adjacent predecessors of block ¥© counts are defined via an enumeration process and over all training sentence pairs. The identity of is ig- ¥ ¨ *) ¨ ¡ are used to define the orientation model : ¡ -' nored. is the number of times the block succeeds some right adjacent predecessor block ¥© . The ’right’ ori- ,+ ¨ ,/ ¨ entation count ¡ is defined accordingly. Note, that ,) *+ ¥ £¦* ¡ ¨>= ¨%0 ¨ ¨ ! . ! ¡ ¡ ¡ *) *+ ¡ ¨ 2 ¨10 in general the unigram count : (' ¡ ¡ during enumeration, a block might have both left and & Trigram language model: The block language right adjacent predecessors, either a left or a right adja- ¡ ¨ ¦ cent predecessor, or no adjacent predecessors at all. The model score ¡ is computed as the proba- orientation count collection is illustrated in Fig. 2: each bility of the first target word in the target clump of given the final two words of the target clump of time a block has a left or right adjacent predecessor in ¡ . the parallel training data, the orientation counts are incre- mented accordingly. The three models are combined in a log-linear way, as The decoding orientation restrictions are illustrated in ¥ shown in the following section. Fig 3: a monotone block sequence with right ( ! ) 1 2 Orientation Unigram Model We keep all blocks for which ?A@CBED¢FHG and the phrase length is less or equal I . No other selection criteria are applied. ?A@CBEDLF The basic idea of the orientation model can be illustrated For the JLK%MONQP model, we keep all blocks for which ¥ as follows: In the example translation in Fig. 1, block R . N (b) += 1 N (b) += 1 L R b4 b4 o =R o =N 4 b3 4 b3 b b o =R o =L 3 b2 3 o =R o =N b' b' 2 b1 2 b1 b2 o =R o =R 1 1 Figure 2: During training, blocks are enumerated in some Figure 3: During decoding, a mostly monotone block se- ¥ ¨ order: for each block , we look for left and right adjacent Q! quence with ¡ orientation is generated as shown © predecessors . in the left picture. In the right picture, block swapping generates block to the left of block . The blocks orientation is generated. If a block is skipped e.g. block and ¡ do not have a left or right adjacent predecessor. ¨ ¨ in Fig 3 by first generating block then block , the ¨ ¥ ! ¤£ £¦¥§£ ¡©¨ block is generated using left orientation . Since ¡ the block translation is generated from bottom-to-top, the over all block segmentations with orientation ¡ for ¨ ¡ which the source phrases yield a segmentation of the in- £¦ ¥© blocks and do not have adjacent predecessors below ¨ ¡ ¨ ¡ ¦ put sentence. Swapping involves only blocks for ¡ *) them: they are generated by a default model ¨ which ¡ for the successor block , e.g. the ¨§ without orientation component. The orientation model ¨ blocks and in Fig 1. We tried several thresholds for *) is given in Eq. 2, the default model is given in Eq. 3. ¨ £¦¥ £¦* ¡¥£¦¥ ¡ ¨ ¡ #. , and performance is reduced significantly only if ¡ *) The block bigram model in ¨ 7 ¡ . No other parameters are used to control © Eq. 1 is defined as: ¥ the block swapping. In particular the orientation of the £¦¥ £¦* ¡ £¦¥ ¡ © ¨ # ! ¡ (2) predecessor block is ignored: in future work, we might © ¤ ¡ ¤ ¥ £ ¨£¢¥¤ ¨£¢§¦ ¨£¢¥ ! take into account that a certain predecessor block typi- ¡ ¡ ¡ ('©¨ cally precedes other blocks. ¡ ¥ ¡ < 0 0 7 ! where 2 and the orientation of the predecessor is ignored. The are chosen to be optimal 3 Experimental Results on the devtest set (the optimal parameter setting is shown in Table.

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    4 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us