Heafield, K. and A. Lavie. "Voting on N-Grams for Machine Translation

Total Page:16

File Type:pdf, Size:1020Kb

Heafield, K. and A. Lavie. Voting on N-grams for Machine Translation System Combination Kenneth Heafield and Alon Lavie Language Technologies Institute Carnegie Mellon University 5000 Forbes Avenue Pittsburgh, PA 15213 fheafield,[email protected] Abstract by a single weight and further current features only guide decisions at the word level. To remedy this System combination exploits differences be- situation, we propose new features that account for tween machine translation systems to form a multiword behavior, each with a separate set of sys- combined translation from several system out- tem weights. puts. Core to this process are features that reward n-gram matches between a candidate 2 Features combination and each system output. Systems differ in performance at the n-gram level de- Most combination schemes generate many hypothe- spite similar overall scores. We therefore ad- sis combinations, score them using a battery of fea- vocate a new feature formulation: for each tures, and search for a hypothesis with highest score system and each small n, a feature counts to output. Formally, the system generates hypoth- n-gram matches between the system and can- esis h, evaluates feature function f, and multiplies didate. We show post-evaluation improvement of 6.67 BLEU over the best system on NIST linear weight vector λ by feature vector f(h) to ob- T MT09 Arabic-English test data. Compared to tain score λ f(h). The score is used to rank fi- a baseline system combination scheme from nal hypotheses and to prune partial hypotheses dur- WMT 2009, we show improvement in the ing beam search. The beams contain hypotheses of range of 1 BLEU point. equal length. With the aim of improving score and therefore translation quality, this paper focuses on the structure of features f and their corresponding 1 Introduction weights λ. System combination merges the output of several The feature function f consists of the following machine translation systems to form an improved feature categories: translation. While individual systems perform sim- Length Length of the hypothesis, as in Koehn et al. ilarly overall, human evaluators report different er- (2007). This compensates, to first order, for the ror distributions for each system (Peterson et al., impact of length on other features. 2009b). For example, some systems are weak at word order while others have more trouble with LM Log probability from an SRI (Stolcke, 2002) nouns and verbs. Existing system combination tech- language model. When the language model niques (Rosti et al., 2008; Karakos et al., 2008; scores a word, it finds the longest n-gram in the Leusch et al., 2009a) ignore these distinctions by model with the same word and context. We use learning a single weight for each system. This the length n as a second feature. The purpose weight is used in word-level decisions and therefore of this second feature is to provide the scor- captures only lexical choice. We see two problems: ing model with limited control over language system behavior differs in more ways than captured model backoff penalties. System 1: Supported Proposal of France tem and across related systems. Viewed as language modeling, each cs;n is a miniature language model System 2: Support for the Proposal of France trained on the translated sentence output by system s and jointly interpolated with a traditional language Candidate: Support for Proposal of France model and with peer models. As described further in Section 4, we jointly tune Unigram Bigram Trigram the weights λ using modified minimum error rate System 1 4 2 1 training (Och, 2003). In doing so, we simultane- System 2 5 3 1 ously learn several weights for each system, one for each length n-gram output by that system. The un- Figure 1: Example match feature values with two sys- igram weight captures confidence in lexical choice tems and matches up to length l = 3. Here, “Supported” while weights on longer n-gram features capture counts because it aligns with “Support”. confidence in word order and phrasal choices. These features are most effective with a variety of hypothe- Match The main focus of this paper, this feature ses from which to choose, so in Section 5 we de- category captures how well the hypothesis cor- scribe a search space with more flexible word order. responds to the system outputs. It consists of Combined, these comprise a combination scheme several features of the form cs;n which count that differs from others in three key ways: the search n-grams in the hypothesis that match the sen- space is more flexible, the features consider matches tence output by system s. We use several in- longer than unigrams, and system weights differ by stantiations of feature cs;n, one for each system task. s and for length n ranging from 1 to some max- imum length l. The hyper parameter l is con- 3 Related Work sidered in Section 6.1. Figure 1 shows example values with two systems and l = 3. The num- System combination takes a variety of forms that ber of match features is l times the number of pair a space of hypothesis combinations with fea- systems being combined. tures to score these hypotheses. Here, we are pri- marily interested in three aspects of each combina- The match features consider only the specific tion scheme: the space of hypotheses, features that sentences being combined, all of which are reward n-gram matches with system outputs, and the translations from a single source sentence. By system weights used for those features. contrast, Leusch et al. (2009a) count matches to the entire translated corpus. While our ap- 3.1 Search Spaces proach considers the most specific data avail- Hypothesis selection (Hildebrand and Vogel, 2009) able, theirs has the advantage of gathering and minimum Bayes risk (Kumar and Byrne, 2004) more data with which to compare, including select from k-best lists output by each system. In document-level matches. As discussed in Sec- the limit case for large k, DeNero et al. (2010) adopt tion 3, we also differ significantly from Leusch the search spaces of the translation systems being et al. (2009a) in that our features have tunable combined. system and n-gram weights while theirs are hy- per parameters. Confusion networks preserve the word order of one k-best list entry called the backbone. The back- Together, these define the feature function f. We bone is chosen by hypothesis selection (Karakos et focus primarily on the new match features that cap- al., 2008; Sim et al., 2007) or jointly with decod- ture lexical and multiword agreement with each sys- ing (Leusch et al., 2009a; Rosti et al., 2008). Other tem. The weight on match count cs;n corresponds k-best entries are aligned to the backbone. The to confidence in n-grams from system s. However, search space consists of choosing each word from this weight also accounts for correlation between among the alternatives aligned with it, keeping these features, which is quite high within the same sys- in backbone order. With this search space, the im- pact of our match features is limited to selection at system weights as a hyper parameter (Heafield et the word level, multiword lexical choice, and possi- al., 2009; Hildebrand and Vogel, 2009). The hy- bly selection of the backbone. per parameter might be set to an increasing function Flexible ordering schemes use a reordering model of each system’s overall performance (Rosti et al., (He and Toutanova, 2009) or dynamically switch 2009; Zhao and Jiang, 2009); this does not account backbones (Heafield et al., 2009) to create word or- for correlation. ders not seen in any single translation. For example, Some combination schemes (Karakos et al., 2008; these translations appear in NIST MT09 (Peterson Leusch et al., 2009a; Rosti et al., 2008; Zens and et al., 2009a): “We have not the interest of control Ney, 2006) tune system weights, but only as they on the Palestinians life,” and “We do not have a de- apply to votes on unigrams. In particular, Leusch sire to control the lives of the Palestinians.” Flexible et al. (2009a) note that their system weights impact ordering schemes consider “have not” versus “do neither their trigram language model based on sys- not have” separately from “Palestinians life” ver- tem outputs nor selection of a backbone. We answer sus “lives of the Palestinians.” Our match features these issues by introducing separate system weights have the most impact here because word order is less for each n-gram length in the system output lan- constrained. This is the type of search space we use guage model and by not restricting search to the or- in our experiments. der of a single backbone. Our experiments use only 1-best outputs, so 3.2 N-gram Match Features system-level weights suffice here. For methods that Agreement is central to system combination and use k-best lists, system weight may be moderated most schemes have some form of n-gram match fea- by some decreasing function of rank in the k-best tures. Of these, the simplest consider only unigram list (Ayan et al., 2008; Zhao and He, 2009). Mini- matches (Ayan et al., 2008; Heafield et al., 2009; mum Bayes risk (Kumar and Byrne, 2004) takes the Rosti et al., 2008; Zhao and Jiang, 2009). technique a step further by using the overall system Some schemes go beyond unigrams but with fixed scores that determined the ranking. weight. Karakos (2009) uses n-gram matches to se- lect the backbone but only unigrams for decoding. 4 Parameter Tuning Kumar and Byrne (2004) use arbitrary evaluation metric to measure similarity.
Recommended publications
  • Integrating Optical Character Recognition and Machine
    Integrating Optical Character Recognition and Machine Translation of Historical Documents Haithem Afli and Andy Way ADAPT Centre School of Computing Dublin City University Dublin, Ireland haithem.afli, andy.way @adaptcentre.ie { } Abstract Machine Translation (MT) plays a critical role in expanding capacity in the translation industry. However, many valuable documents, including digital documents, are encoded in non-accessible formats for machine processing (e.g., Historical or Legal documents). Such documents must be passed through a process of Optical Character Recognition (OCR) to render the text suitable for MT. No matter how good the OCR is, this process introduces recognition errors, which often renders MT ineffective. In this paper, we propose a new OCR to MT framework based on adding a new OCR error correction module to enhance the overall quality of translation. Experimenta- tion shows that our new system correction based on the combination of Language Modeling and Translation methods outperforms the baseline system by nearly 30% relative improvement. 1 Introduction While research on improving Optical Character Recognition (OCR) algorithms is ongoing, our assess- ment is that Machine Translation (MT) will continue to produce unacceptable translation errors (or non- translations) based solely on the automatic output of OCR systems. The problem comes from the fact that current OCR and Machine Translation systems are commercially distinct and separate technologies. There are often mistakes in the scanned texts as the OCR system occasionally misrecognizes letters and falsely identifies scanned text, leading to misspellings and linguistic errors in the output text (Niklas, 2010). Works involved in improving translation services purchase off-the-shelf OCR technology but have limited capability to adapt the OCR processing to improve overall machine translation perfor- mance.
    [Show full text]
  • Automatic Correction of Real-Word Errors in Spanish Clinical Texts
    sensors Article Automatic Correction of Real-Word Errors in Spanish Clinical Texts Daniel Bravo-Candel 1,Jésica López-Hernández 1, José Antonio García-Díaz 1 , Fernando Molina-Molina 2 and Francisco García-Sánchez 1,* 1 Department of Informatics and Systems, Faculty of Computer Science, Campus de Espinardo, University of Murcia, 30100 Murcia, Spain; [email protected] (D.B.-C.); [email protected] (J.L.-H.); [email protected] (J.A.G.-D.) 2 VÓCALI Sistemas Inteligentes S.L., 30100 Murcia, Spain; [email protected] * Correspondence: [email protected]; Tel.: +34-86888-8107 Abstract: Real-word errors are characterized by being actual terms in the dictionary. By providing context, real-word errors are detected. Traditional methods to detect and correct such errors are mostly based on counting the frequency of short word sequences in a corpus. Then, the probability of a word being a real-word error is computed. On the other hand, state-of-the-art approaches make use of deep learning models to learn context by extracting semantic features from text. In this work, a deep learning model were implemented for correcting real-word errors in clinical text. Specifically, a Seq2seq Neural Machine Translation Model mapped erroneous sentences to correct them. For that, different types of error were generated in correct sentences by using rules. Different Seq2seq models were trained and evaluated on two corpora: the Wikicorpus and a collection of three clinical datasets. The medicine corpus was much smaller than the Wikicorpus due to privacy issues when dealing Citation: Bravo-Candel, D.; López-Hernández, J.; García-Díaz, with patient information.
    [Show full text]
  • Machine Translation for Academic Purposes Grace Hui-Chin Lin
    Proceedings of the International Conference on TESOL and Translation 2009 December 2009: pp.133-148 Machine Translation for Academic Purposes Grace Hui-chin Lin PhD Texas A&M University College Station Master of Science, University of Southern California Paul Shih Chieh Chien Center for General Education, Taipei Medical University Abstract Due to the globalization trend and knowledge boost in the second millennium, multi-lingual translation has become a noteworthy issue. For the purposes of learning knowledge in academic fields, Machine Translation (MT) should be noticed not only academically but also practically. MT should be informed to the translating learners because it is a valuable approach to apply by professional translators for diverse professional fields. For learning translating skills and finding a way to learn and teach through bi-lingual/multilingual translating functions in software, machine translation is an ideal approach that translation trainers, translation learners, and professional translators should be familiar with. In fact, theories for machine translation and computer assistance had been highly valued by many scholars. (e.g., Hutchines, 2003; Thriveni, 2002) Based on MIT’s Open Courseware into Chinese that Lee, Lin and Bonk (2007) have introduced, this paper demonstrates how MT can be efficiently applied as a superior way of teaching and learning. This article predicts the translated courses utilizing MT for residents of global village should emerge and be provided soon in industrialized nations and it exhibits an overview about what the current developmental status of MT is, why the MT should be fully applied for academic purpose, such as translating a textbook or teaching and learning a course, and what types of software can be successfully applied.
    [Show full text]
  • Machine Translation for Language Preservation
    Machine translation for language preservation Steven BIRD1,2 David CHIANG3 (1) Department of Computing and Information Systems, University of Melbourne (2) Linguistic Data Consortium, University of Pennsylvania (3) Information Sciences Institute, University of Southern California [email protected], [email protected] ABSTRACT Statistical machine translation has been remarkably successful for the world’s well-resourced languages, and much effort is focussed on creating and exploiting rich resources such as treebanks and wordnets. Machine translation can also support the urgent task of document- ing the world’s endangered languages. The primary object of statistical translation models, bilingual aligned text, closely coincides with interlinear text, the primary artefact collected in documentary linguistics. It ought to be possible to exploit this similarity in order to improve the quantity and quality of documentation for a language. Yet there are many technical and logistical problems to be addressed, starting with the problem that – for most of the languages in question – no texts or lexicons exist. In this position paper, we examine these challenges, and report on a data collection effort involving 15 endangered languages spoken in the highlands of Papua New Guinea. KEYWORDS: endangered languages, documentary linguistics, language resources, bilingual texts, comparative lexicons. Proceedings of COLING 2012: Posters, pages 125–134, COLING 2012, Mumbai, December 2012. 125 1 Introduction Most of the world’s 6800 languages are relatively unstudied, even though they are no less im- portant for scientific investigation than major world languages. For example, before Hixkaryana (Carib, Brazil) was discovered to have object-verb-subject word order, it was assumed that this word order was not possible in a human language, and that some principle of universal grammar must exist to account for this systematic gap (Derbyshire, 1977).
    [Show full text]
  • Unified Language Model Pre-Training for Natural
    Unified Language Model Pre-training for Natural Language Understanding and Generation Li Dong∗ Nan Yang∗ Wenhui Wang∗ Furu Wei∗† Xiaodong Liu Yu Wang Jianfeng Gao Ming Zhou Hsiao-Wuen Hon Microsoft Research {lidong1,nanya,wenwan,fuwei}@microsoft.com {xiaodl,yuwan,jfgao,mingzhou,hon}@microsoft.com Abstract This paper presents a new UNIfied pre-trained Language Model (UNILM) that can be fine-tuned for both natural language understanding and generation tasks. The model is pre-trained using three types of language modeling tasks: unidirec- tional, bidirectional, and sequence-to-sequence prediction. The unified modeling is achieved by employing a shared Transformer network and utilizing specific self-attention masks to control what context the prediction conditions on. UNILM compares favorably with BERT on the GLUE benchmark, and the SQuAD 2.0 and CoQA question answering tasks. Moreover, UNILM achieves new state-of- the-art results on five natural language generation datasets, including improving the CNN/DailyMail abstractive summarization ROUGE-L to 40.51 (2.04 absolute improvement), the Gigaword abstractive summarization ROUGE-L to 35.75 (0.86 absolute improvement), the CoQA generative question answering F1 score to 82.5 (37.1 absolute improvement), the SQuAD question generation BLEU-4 to 22.12 (3.75 absolute improvement), and the DSTC7 document-grounded dialog response generation NIST-4 to 2.67 (human performance is 2.65). The code and pre-trained models are available at https://github.com/microsoft/unilm. 1 Introduction Language model (LM) pre-training has substantially advanced the state of the art across a variety of natural language processing tasks [8, 29, 19, 31, 9, 1].
    [Show full text]
  • Evaluation of Machine Learning Algorithms for Sms Spam Filtering
    EVALUATION OF MACHINE LEARNING ALGORITHMS FOR SMS SPAM FILTERING David Bäckman Bachelor Thesis, 15 credits Bachelor Of Science Programme in Computing Science 2019 Abstract The purpose of this thesis is to evaluate dierent machine learning algorithms and methods for text representation in order to determine what is best suited to use to distinguish between spam SMS and legitimate SMS. A data set that contains 5573 real SMS has been used to train the algorithms K-Nearest Neighbor, Support Vector Machine, Naive Bayes and Logistic Regression. The dierent methods that have been used to represent text are Bag of Words, Bigram and Word2Vec. In particular, it has been investigated if semantic text representations can improve the performance of classication. A total of 12 combinations have been evaluated with help of the metrics accuracy and F1-score. The results shows that Logistic Regression together with Bag of Words reach the highest accuracy and F1-score. Bigram as text representation seems to work worse then the others methods. Word2Vec can increase the performnce for K- Nearst Neigbor but not for the other algorithms. Acknowledgements I would like to thank my supervisor Kai-Florian Richter for all good advice and guidance through the project. I would also like to thank all my classmates for help and support during the education, you have made it possible for me to reach this day. Contents 1 Introduction 1 1.1 Background1 1.2 Purpose and Research Questions1 2 Related Work 3 3 Theoretical Background 5 3.1 The Choice of Algorithms5 3.2 Classication
    [Show full text]
  • Machine Translation and Computer-Assisted Translation: a New Way of Translating? Author: Olivia Craciunescu E-Mail: Olivia [email protected]
    Machine Translation and Computer-Assisted Translation: a New Way of Translating? Author: Olivia Craciunescu E-mail: [email protected] Author: Constanza Gerding-Salas E-mail: [email protected] Author: Susan Stringer-O’Keeffe E-mail: [email protected] Source: http://www.translationdirectory.com The Need for Translation IT has produced a screen culture that Technology tends to replace the print culture, with printed documents being dispensed Advances in information technology (IT) with and information being accessed have combined with modern communication and relayed directly through computers requirements to foster translation automation. (e-mail, databases and other stored The history of the relationship between information). These computer documents technology and translation goes back to are instantly available and can be opened the beginnings of the Cold War, as in the and processed with far greater fl exibility 1950s competition between the United than printed matter, with the result that the States and the Soviet Union was so intensive status of information itself has changed, at every level that thousands of documents becoming either temporary or permanent were translated from Russian to English and according to need. Over the last two vice versa. However, such high demand decades we have witnessed the enormous revealed the ineffi ciency of the translation growth of information technology with process, above all in specialized areas of the accompanying advantages of speed, knowledge, increasing interest in the idea visual
    [Show full text]
  • NLP - Assignment 2
    NLP - Assignment 2 Week 2 December 27th, 2016 1. A 5-gram model is a order Markov Model: (a) Six (b) Five (c) Four (d) Constant Ans : c) Four 2. For the following corpus C1 of 3 sentences, what is the total count of unique bi- grams for which the likelihood will be estimated? Assume we do not perform any pre-processing, and we are using the corpus as given. (i) ice cream tastes better than any other food (ii) ice cream is generally served after the meal (iii) many of us have happy childhood memories linked to ice cream (a) 22 (b) 27 (c) 30 (d) 34 Ans : b) 27 3. Arrange the words \curry, oil and tea" in descending order, based on the frequency of their occurrence in the Google Books n-grams. The Google Books n-gram viewer is available at https://books.google.com/ngrams: (a) tea, oil, curry (c) curry, tea, oil (b) curry, oil, tea (d) oil, tea, curry Ans: d) oil, tea, curry 4. Given a corpus C2, The Maximum Likelihood Estimation (MLE) for the bigram \ice cream" is 0.4 and the count of occurrence of the word \ice" is 310. The likelihood of \ice cream" after applying add-one smoothing is 0:025, for the same corpus C2. What is the vocabulary size of C2: 1 (a) 4390 (b) 4690 (c) 5270 (d) 5550 Ans: b)4690 The Questions from 5 to 10 require you to analyse the data given in the corpus C3, using a programming language of your choice.
    [Show full text]
  • Language Service Translation SAVER Technote
    TechNote August 2020 LANGUAGE TRANSLATION APPLICATIONS The U.S. Department of Homeland Security (DHS) AEL Number: 09ME-07-TRAN established the System Assessment and Validation for Language translation equipment used by emergency services has evolved into Emergency Responders commercially available, mobile device applications (apps). The wide adoption of (SAVER) Program to help cell phones has enabled language translation applications to be useful in emergency responders emergency scenarios where first responders may interact with individuals who improve their procurement speak a different language. Most language translation applications can identify decisions. a spoken language and translate in real-time. Applications vary in functionality, Located within the Science such as the ability to translate voice or text offline and are unique in the number and Technology Directorate, of available translatable languages. These applications are useful tools to the National Urban Security remedy communication barriers in any emergency response scenario. Technology Laboratory (NUSTL) manages the SAVER Overview Program and conducts objective operational During an emergency response, language barriers can challenge a first assessments of commercial responder’s ability to quickly assess and respond to a situation. First responders equipment and systems must be able to accurately communicate with persons at an emergency scene in relevant to the emergency order to provide a prompt, appropriate response. The quick translations provided responder community. by mobile apps remove the language barrier and are essential to the safety of The SAVER Program gathers both first responders and civilians with whom they interact. and reports information about equipment that falls within the Before the adoption of smartphones, first responders used language interpreters categories listed in the DHS and later language translation equipment such as hand-held language Authorized Equipment List translation devices (Figure 1).
    [Show full text]
  • Terminology Extraction, Translation Tools and Comparable Corpora Helena Blancafort, Béatrice Daille, Tatiana Gornostay, Ulrich Heid, Claude Méchoulam, Serge Sharoff
    TTC: Terminology Extraction, Translation Tools and Comparable Corpora Helena Blancafort, Béatrice Daille, Tatiana Gornostay, Ulrich Heid, Claude Méchoulam, Serge Sharoff To cite this version: Helena Blancafort, Béatrice Daille, Tatiana Gornostay, Ulrich Heid, Claude Méchoulam, et al.. TTC: Terminology Extraction, Translation Tools and Comparable Corpora. 14th EURALEX International Congress, Jul 2010, Leeuwarden/Ljouwert, Netherlands. pp.263-268. hal-00819365 HAL Id: hal-00819365 https://hal.archives-ouvertes.fr/hal-00819365 Submitted on 2 May 2013 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. TTC: Terminology Extraction, Translation Tools and Comparable Corpora Helena Blancafort, Syllabs, Universitat Pompeu Fabra, Barcelona, Spain Béatrice Daille, LINA, Université de Nantes, France Tatiana Gornostay, Tilde, Latva Ulrich Heid, IMS, Universität Stuttgart, Germany Claude Mechoulam, Sogitec, France Serge Sharoff, CTS, University of Leeds, England The need for linguistic resources in any natural language application is undeniable. Lexicons and terminologies
    [Show full text]
  • Acknowledging the Needs of Computer-Assisted Translation Tools Users: the Human Perspective in Human-Machine Translation Annemarie Taravella and Alain O
    The Journal of Specialised Translation Issue 19 – January 2013 Acknowledging the needs of computer-assisted translation tools users: the human perspective in human-machine translation AnneMarie Taravella and Alain O. Villeneuve, Université de Sherbrooke, Quebec ABSTRACT The lack of translation specialists poses a problem for the growing translation markets around the world. One of the solutions proposed for the lack of human resources is automated translation tools. In the last few decades, organisations have had the opportunity to increase their use of technological resources. However, there is no consensus on the way that technological resources should be integrated into translation service providers (TSP). The approach taken by this article is to set aside both 100% human translation and 100% machine translation (without human intervention), to examine a third, more realistic solution: interactive translation where humans and machines co-operate. What is the human role? Based on the conceptual framework of information systems and organisational sciences, we recommend giving users, who are mainly translators for whom interactive translation tools are designed, a fundamental role in the thinking surrounding the implementation of a technological tool. KEYWORDS Interactive translation, information system, implementation, business process reengineering, organisational science. 1. Introduction Globalisation and the acceleration of world trade operations have led to an impressive growth of the global translation services market. According to EUATC (European Union of Associations of Translation Companies), translation industry is set to grow around five percent annually in the foreseeable future (Hager, 2008). With its two official languages, Canada is a good example of an important translation market, where highly skilled professional translators serve not only the domestic market, but international clients as well.
    [Show full text]
  • 3 Dictionaries and Tolerant Retrieval
    Online edition (c)2009 Cambridge UP DRAFT! © April 1, 2009 Cambridge University Press. Feedback welcome. 49 Dictionaries and tolerant 3 retrieval In Chapters 1 and 2 we developed the ideas underlying inverted indexes for handling Boolean and proximity queries. Here, we develop techniques that are robust to typographical errors in the query, as well as alternative spellings. In Section 3.1 we develop data structures that help the search for terms in the vocabulary in an inverted index. In Section 3.2 we study WILDCARD QUERY the idea of a wildcard query: a query such as *a*e*i*o*u*, which seeks doc- uments containing any term that includes all the five vowels in sequence. The * symbol indicates any (possibly empty) string of characters. Users pose such queries to a search engine when they are uncertain about how to spell a query term, or seek documents containing variants of a query term; for in- stance, the query automat* would seek documents containing any of the terms automatic, automation and automated. We then turn to other forms of imprecisely posed queries, focusing on spelling errors in Section 3.3. Users make spelling errors either by accident, or because the term they are searching for (e.g., Herman) has no unambiguous spelling in the collection. We detail a number of techniques for correcting spelling errors in queries, one term at a time as well as for an entire string of query terms. Finally, in Section 3.4 we study a method for seeking vo- cabulary terms that are phonetically close to the query term(s).
    [Show full text]