Multi-Dialect Neural Machine Translation and Dialectometry

Multi-Dialect Neural Machine Translation and Dialectometry

PACLIC 32 Multi-dialect Neural Machine Translation and Dialectometry Kaori Abe1, Yuichiroh Matsubayashi2, Naoaki Okazaki3, and Kentaro Inui1;2 1Tohoku University, 2RIKEN, 3Tokyo Institute of Technology [email protected], [email protected] [email protected], [email protected] Abstract take standard Japanese as input could be applied to dialects as well. In addition, if a standard-to-dialect We present a multi-dialect neural machine translation system were also available, smart speak- translation (NMT) model tailored to Japanese. ers could respond to the native speakers of a dialect While the surface forms of Japanese dialects using that dialect. We believe that friendly interac- differ from those of standard Japanese, most tions of this sort might lead to such systems gaining of the dialects share fundamental proper- ties such as word order, and some also use more widespread acceptance of in the Japanese so- many of the same phonetic correspondence ciety. rules. To take advantage of these properties, In this paper, we present a multi-dialect neural we integrate multilingual, syllable-level, and MT (NMT) system tailored to Japanese. Specifi- fixed-order translation techniques into a gen- cally, we employ kana, a Japanese phonetic lettering eral NMT model. Our experimental results system, as basic units in the encoder–decoder frame- demonstrate that this model can outperform a baseline dialect translation model. In addition, work to avoid the following: ambiguity in convert- we show that visualizing the dialect embed- ing kana to kanji (characters in the Japanese writing dings learned by the model can facilitate geo- system); difficulties in identifying word boundaries graphical and typological analyses of dialects. especially for dialects; and data sparseness problems due to handling large numbers of words originating from different dialects. Since dialects almost always 1 Introduction use the same word order as standard Japanese, we Since the use of automated personal assistants (e.g., employ bunsetsu (Japanese phrase units) as a unit of Apple’s Siri, Google Assistant, or Microsoft Cor- sequences rather than sentence which is more com- tana) and smart speakers (e.g., Amazon Alexa or monly used in NMT. Google Home) has become more widespread, de- One issue for Japanese dialects is a lack of train- mand has also grown to bridge the gap between di- ing data. To deal with this, we build a unified alects and the standard form of a given language. NMT model covering multiple dialects, inspired by The importance of handling dialects is especially ev- work on multilingual NMT (Johnson et al., 2016). ident in a rapidly aging society like Japan, where This approach utilizes dialect embeddings, namely older people use them extensively. vector representations of Japanese dialects, to in- To address this issue, we consider a system for form the model of the input dialect. An interesting machine translation (MT) between Japanese dialects by-product of this approach is that the dialect em- and standard Japanese. If this can provide cor- beddings the system learns illustrate the difference rect translations in the dialect-to-standard direction, between different dialect types from different geo- then, other natural language processing systems graphical areas. In addition, we present an exam- (e.g., information retrieval or semantic analysis) that ple of using these dialect embeddings for dialectom- 1 32nd Pacific Asia Conference on Language, Information and Computation Hong Kong, 1-3 December 2018 Copyright 2018 by the authors PACLIC 32 etry (Nerbonne and Kretzschmar, 2011; Kumagai, (dialect) source ( = Right, definitely. ) 2016; Guggilla, 2016; Rama and C¸ oltekin,¨ 2016). divided into “bunsetsu” Another advantage of adopting a the multilingual input sequence dialect label ( = Right) dialect label (=definitely) architecture for multiple related languages is it can enable the system to acquire knowledge of their lex- ical and syntactic similarities. For example, Lakew Multi-dialect NMT et al. (2018) reported that including several related languages in supervised training data can improve output sequence Attention multilingual NMT. Our results confirm the effec- tiveness of using closely-related languages (namely concatenate Japanese dialects) in multilingual NMT. (standard JPN) “Hyogo” target 2 Related Work Little dialectal text is available since dialects are Figure 1: Proposed multi-dialect NMT model. generally spoken rather than written. For this reason, many dialect MT researchers work in low-resource multi-dialect model can capture similarities between settings (Zbib et al., 2012; Scherrer and Ljubesiˇ c,´ dialects (Section 5). 2016; Hassan et al., 2017). Previous work reported that character-level sta- However, the use of similar dialects has been tistical machine translation (SMT) using words as found to be helpful in learning translation mod- translation units was effective for translating be- els for particular dialects. Several previous studies tween closely-related languages (Nakov and Tiede- have investigated the characteristics of translation mann, 2012; Scherrer and Ljubesiˇ c,´ 2016). There models for closely-related dialects (Meftouh et al., are two reasons for this: the character-level informa- 2015; Honnet et al., 2018). For example, Honnet tion enables the system to exploit lexical overlaps, et al. (2018) reported that a character-level NMT while using words as translation units takes advan- model trained on one Swiss-German dialect per- tage of related languages’ syntactic overlaps. formed moderately well at translating sentences in In this study, we present a method of translat- closely-related dialects. ing between Japanese dialects that combines three Therefore, given this, we use multilingual ideas: multilingual NMT, character-level NMT, and NMT (Johnson et al., 2016) to learn parameters the use of base phrases (i.e., bunsetsu) as translation that encode knowledge of dialects’ shared lexical units. We believe this enables our approach to fully and syntactic structure. Gu et al. (2018) demon- exploit the similarities among dialects and standard strated that multilingual NMT can be useful for low- Japanese, even in low-resource settings. resource language pairs, while Lakew et al. (2018) found that a multilingual NMT system trained on 3 Data: Japanese Dialect Corpus multiple related languages showed improved zero- shot translation performance. We believe that multi- Japanese is a dialect-rich language, with dozens of lingual NMT can be effective for closely-related di- dialects that are used for everyday conversations in alects, and can compensate for a lack of translation most Japanese regions. They can be characterized data for the different dialects. in terms of differences in their content words (vo- Multilingual NMT can also help us to analyze the cabulary) and regular phonetic shifts, mostly in their characteristics of each language. Ostling¨ and Tiede- postpositions and suffixes. That said, they nonethe- mann (2016) found that clustering the language em- less share most words with standard Japanese, as beddings learned by a character-level multilingual well as mostly using common grammatical rules, system provided an illustration of the language fam- such as for word ordering, syntactic marker cate- ilies involved. In the light of this, we also analyze gories, and connecting syntactic markers. Some di- our dialect embeddings to investigate whether our alects also share the dialect-specific vocabulary. 2 32nd Pacific Asia Conference on Language, Information and Computation Hong Kong, 1-3 December 2018 Copyright 2018 by the authors PACLIC 32 Figure 2: Number of sentences in each dialect in The National Dialect Discourse Database Collection. In this study, we used a collection of parallel tex- ID Encoder input order tual data for dialects and standard Japanese, the Na- a source label, sentence tional Dialect Discourse Database Collection (NIN- b target label, sentence JAL, 1980). This corpus includes 48 dialects, one c source label, target label, sentence from each of the 47 prefectures and an additional di- d source label, sentence, target label alect from the Okinawa Prefecture. For each dialect, the texts consist of transcribed 30-minute conversa- Table 1: Dialect label input order variants. tions between two native speakers of that dialect. The total number of dialect sentences (each paired using dialect embeddings that included auxiliary di- with a translation into standard Japanese) is 34,117, alect labels. Johnson et al. (2016) added a label to and Figure 2 shows the number of sentences in each the beginning of each sequence to specify the output dialect. language. We modify this to specify both the input Japanese texts are generally written in a mix of and output dialects to the model, and examine the kanji and kana; therefore, we converted the kanji in four different placements for these labels as shown the sentences into kana and, then, segmented them in Table 1. into bunsetsu.1 After preprocessing, the average sentence lengths were 14.62 and 15.57 characters for Syllable-to-syllable translation As mentioned in the dialects and standard Japanese, respectively, and Section 3, one of the keys to translating between two the average number of bunsetsus per sentence was closely-related languages is to model the phonetic 3.42. correspondences between them. Thus, to consider syllable-level translation rules that may be shared 4 NMT Model by similar dialects,

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    10 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us