Dictionary Look-Up Within Small Edit Distance

Total Page:16

File Type:pdf, Size:1020Kb

Dictionary Look-Up Within Small Edit Distance Dictionary Lo okUp Within Small Edit Distance Ab dullah N Arslan and Omer Egeciog lu Department of Computer Science University of California Santa Barbara Santa Barbara CA USA farslanomergcsucsbedu Abstract Let W b e a dictionary consisting of n binary strings of length m each represented as a trie The usual dquery asks if there exists a string in W within Hamming distance d of a given binary query string q We present an algorithm to determine if there is a memb er in W within edit distance d of a given query string q of length m The metho d d+1 takes time O dm in the RAM mo del indep endent of n and requires O dm additional space Intro duction Let W b e a dictionary consisting of n binary strings of length m each A dquery asks if there exists a string in W within Hamming distance d of a given binary query string q Algorithms for answering dqueries eciently has b een a topic of interest for some time and have also b een studied as the approximate query and the approximate query retrieval problems in the literature The problem was originally p osed by Minsky and Pap ert in in which they asked if there is a data structure that supp orts fast dqueries The cases of small d and large d for this problem seem to require dierent techniques for their solutions The case when d is small was studied by Yao and Yao Dolev et al and Greene et al have made some progress when d is relatively large There are ecient algorithms only when d prop osed by Bro dal and Venkadesh Yao and Yao and Bro dal and Gasieniec The small d case has applications in password security Searching biological sequence databases may also use the metho ds of answering dqueries Previous studies for the dquery problem have fo cused on minimizing the numb er of memory accesses for a dquery assuming other computations are free and used cell or bit prob e mo dels to express complexity We assume a RAM mo del with constant memory access time and take into account all computations in the complexity analysis Dolev et al presented b ounds for the space and time complexity of the dquery problem under certain assumptions using various notions of proximity In the mo del W is stored in buckets and prepro cessing of W is allowed Supp orted in part by NSF Grant No EIA In this pap er we consider answering dqueries eciently without limiting our selves to the construction of a new data structure parametrized by d The variant of the original dquery problem that we consider is when the stringtostring edit distance is used as the distance measure instead of the ordinary case of Hamming distance We assume that W is stored as a trie T and prop ose two algorithms m for the dquery problem in this case Our algorithms use the hybrid treedynamic programming approach The rst one Algorithm LOOKUP Figure re ed d+2 d+1 quires O dm time in the worst case and O dm space in addition to the space requirements of the trie T This complexity is of interest for small values m of d under investigation The second algorithm Algorithm DFTLOOKUP ed d+1 Figure has time complexity O dm and additional space complexity of only O dm There is reason to b elieve that the average p erformance of b oth algorithms is much b etter when W is sparse Motivation Hamming Distance Based Metho ds Hamming distance b etween two binary strings is the numb er of p ositions they dier A dquery asks if there is a memb er in a dictionary W whose Hamming distance is at most d from a given binary query string q We assume a trie representation T for W and assume for simplicity that m W consists of binary words of length m each A trie is a tree whose arcs are lab eled by the symb ols of alphab et in this case f g The leaf no des of T corresp ond to the words in W and when concatenated the lab els of arcs on m a path from the ro ot to a given intermediate no de gives a prex of at least one word in W Clearly in the RAM mo del assumed accessing a word in W takes O m time Figure part a shows an example trie T representing a dictionary 5 W f g Root Root 0 0 1 0 1 0 1 0 1 1 0 1 1 0 1 2 0 0 1 0 0 1 1 2 2 1 0 1 1 0 1 2 2 3 1 1 1 1 1 1 3 3 4 (a) (b) Fig a An example trie with binary words and b The numb ers in italic are the no de weights computed with resp ect to query string A naive metho d for answering a dquery is to generate the whole set of P d m strings diering from q in at most d p ositions and with every string k =0 k generated p erform a dictionary lo okup in T for an exact memb er in W This m d+2 naive generate and test algorithm takes O m time and O m additional space to store a generated string at a time Another naive metho d is to add all strings within Hamming distance d from any memb er in W to obtain a bigger dictionary W Then any dquery can b e answered in O m time using the corresp onding trie T for an exact memb er This latter metho d signicantly increases the size m d of W by a numb er roughly O nm mbit memb ers Cost of constructing and maintaining T may b e extremely high m For Hamming distance we can improve the rst naive algorithm ab ove as follows Let sv denote the prex corresp onding to trie no de v Given a query string q supp ose that we assign weight w to each trie no de v in T as h m w v hsv q h 1js(v )j where h denotes Hamming distance As an example in Figure b the weights of the no des have b een computed with resp ect to query string q The idea is that we can prune the trie in our search for q at the no des in T with m w v d h Lemma Let N be the number of nodes in T with weight d as dened in m d+1 Then N O m Proof It is easy to see that N is maximized over all tries T when T is a m m complete trie over ie T contains all binary strings of length m Figure m shows no de weights of a complete trie with resp ect to q up to level starting with the ro ot at level The ro ot has weight For any other vertex v at level l if the arc from its parent to v has lab el q the l th symb ol of query string q l then v and its parent have the same weight otherwise weight of v is more than that of its parent Let Ll w denote the numb er of vertices with weight w at level l of the complete trie T At any level l the largest weight is l Using m these observations we see that Ll w Ll w Ll w with l w and l Ll Therefore Ll w is the binomial co ecient Furthermore since w the smallest level at which weight w app ears in T is l w the total numb er m w w +1 m m+1 of vertices with weight w in T is Hence m w w w w +1 d X m d+1 O m N w w =0 Based on the ab ove lemma Figure outlines Algorithm LOOKUP for h dictionary lo okup within Hamming distance d The algorithm explores all no des v in T with weight w v d ie sv is a prex of a word in W whose m h Hamming distance from q is p otentially within d S stores the set of no de k weight pairs v w v for all no des v at levels k with weight w v d The h h algorithm iteratively computes S from S by collecting all pairs v a w v k k 1 ha q in S where v w v S w v ha q d and a f g k k k 1 k Level 0 0 0 1 0 1 1 0 1 0 1 0 1 1 2 2 0 1 0 1 0 1 0 1 0 1 1 2 1 2 2 3 3 0 10 10 10 10 10 10 10 1 4 01121223122323 34 Fig In a complete binary trie of height weights with resp ect to a given binary query string are shown in italic Clearly if there is a memb er in W within Hamming distance d then it will b e captured in S in which case the algorithm returns YES otherwise it returns m NO d+1 S contains O m no deweight pairs by Lemma Therefore the time k d+2 complexity in the assumed mo del is O m It also requires additional space d+1 to store O m trie no des The time complexity is no b etter in the worst case than that of the naive algorithm which generates and tries all p ossible strings within Hamming distance d from q However for a sparse dictionary W Algorithm LOOKUP is b ound to b e much faster on the average h Algorithm LOOKUP d h S fv g 0 0 for k to m do S fv a w v ha q j v w v S w v ha q d and a g k k k 1 k Return YES if the minimum weight in S is d NO otherwise m Fig Algorithm LOOKUP for dictionary lo okup within Hamming distance d h Edit Distance Based dqueries Algorithm LOOKUP essentially computes the Hamming distance b etween q h and a selected part of dictionary W Next we investigate the p ossibility of using a similar idea for dqueries dened with resp ect to the edit distance For the purp oses of this pap er we use a simple typ e of edit distance Given two strings p p p and q q q the edit distance edp q is the 1 m 1 m minimum numb er of edit op erations which transforms p into q The edit op er ations are of three typ es insert delete and substitute Substituting a symb ol by itself is called a match A match op eration is the only op eration that do es not contribute to the numb er of steps of the transformation In terms of costs all edit op erations have cost except for the match whose cost is The usual framework for the analysis of edit distance is the edit graph Edit Graph G pq 2 is a directed acyclic graph having m lattice p oints u v as vertices for u v m Horizontal and vertical arcs corresp ond to insert and delete op er ations resp ectively The diagonal arcs corresp ond to substitutions Each arc has a cost corresp onding to the edit op eration it represents If we trace the arcs of a path from no de to an intermediate no de i j and p erform the indicated edit op erations in the given order on p p then we obtain q q Edit dis 1 i 1 j tance b etween prexes p p and q q is the cost of the minimumcost
Recommended publications
  • The One-Way Communication Complexity of Hamming Distance
    THEORY OF COMPUTING, Volume 4 (2008), pp. 129–135 http://theoryofcomputing.org The One-Way Communication Complexity of Hamming Distance T. S. Jayram Ravi Kumar∗ D. Sivakumar∗ Received: January 23, 2008; revised: September 12, 2008; published: October 11, 2008. Abstract: Consider the following version of the Hamming distance problem for ±1 vec- n √ n √ tors of length n: the promise is that the distance is either at least 2 + n or at most 2 − n, and the goal is to find out which of these two cases occurs. Woodruff (Proc. ACM-SIAM Symposium on Discrete Algorithms, 2004) gave a linear lower bound for the randomized one-way communication complexity of this problem. In this note we give a simple proof of this result. Our proof uses a simple reduction from the indexing problem and avoids the VC-dimension arguments used in the previous paper. As shown by Woodruff (loc. cit.), this implies an Ω(1/ε2)-space lower bound for approximating frequency moments within a factor 1 + ε in the data stream model. ACM Classification: F.2.2 AMS Classification: 68Q25 Key words and phrases: One-way communication complexity, indexing, Hamming distance 1 Introduction The Hamming distance H(x,y) between two vectors x and y is defined to be the number of positions i such that xi 6= yi. Let GapHD denote the following promise problem: the input consists of ±1 vectors n √ n √ x and y of length n, together with the promise that either H(x,y) ≤ 2 − n or H(x,y) ≥ 2 + n, and the goal is to find out which of these two cases occurs.
    [Show full text]
  • Fault Tolerant Computing CS 530 Information Redundancy: Coding Theory
    Fault Tolerant Computing CS 530 Information redundancy: Coding theory Yashwant K. Malaiya Colorado State University April 3, 2018 1 Information redundancy: Outline • Using a parity bit • Codes & code words • Hamming distance . Error detection capability . Error correction capability • Parity check codes and ECC systems • Cyclic codes . Polynomial division and LFSRs 4/3/2018 Fault Tolerant Computing ©YKM 2 Redundancy at the Bit level • Errors can bits to be flipped during transmission or storage. • An extra parity bit can detect if a bit in the word has flipped. • Some errors an be corrected if there is enough redundancy such that the correct word can be guessed. • Tukey: “bit” 1948 • Hamming codes: 1950s • Teletype, ASCII: 1960: 7+1 Parity bit • Codes are widely used. 304,805 letters in the Torah Redundancy at the Bit level Even/odd parity (1) • Errors can bits to be flipped during transmission/storage. • Even/odd parity: . is basic method for detecting if one bit (or an odd number of bits) has been switched by accident. • Odd parity: . The number of 1-bit must add up to an odd number • Even parity: . The number of 1-bit must add up to an even number Even/odd parity (2) • The it is known which parity it is being used. • If it uses an even parity: . If the number of of 1-bit add up to an odd number then it knows there was an error: • If it uses an odd: . If the number of of 1-bit add up to an even number then it knows there was an error: • However, If an even number of 1-bit is flipped the parity will still be the same.
    [Show full text]
  • Sequence Distance Embeddings
    Sequence Distance Embeddings by Graham Cormode Thesis Submitted to the University of Warwick for the degree of Doctor of Philosophy Computer Science January 2003 Contents List of Figures vi Acknowledgments viii Declarations ix Abstract xi Abbreviations xii Chapter 1 Starting 1 1.1 Sequence Distances . ....................................... 2 1.1.1 Metrics ............................................ 3 1.1.2 Editing Distances ...................................... 3 1.1.3 Embeddings . ....................................... 3 1.2 Sets and Vectors ........................................... 4 1.2.1 Set Difference and Set Union . .............................. 4 1.2.2 Symmetric Difference . .................................. 5 1.2.3 Intersection Size ...................................... 5 1.2.4 Vector Norms . ....................................... 5 1.3 Permutations ............................................ 6 1.3.1 Reversal Distance ...................................... 7 1.3.2 Transposition Distance . .................................. 7 1.3.3 Swap Distance ....................................... 8 1.3.4 Permutation Edit Distance . .............................. 8 1.3.5 Reversals, Indels, Transpositions, Edits (RITE) . .................... 9 1.4 Strings ................................................ 9 1.4.1 Hamming Distance . .................................. 10 1.4.2 Edit Distance . ....................................... 10 1.4.3 Block Edit Distances . .................................. 11 1.5 Sequence Distance Problems . .................................
    [Show full text]
  • Approximate String Matching with Reduced Alphabet
    Approximate String Matching with Reduced Alphabet Leena Salmela1 and Jorma Tarhio2 1 University of Helsinki, Department of Computer Science [email protected] 2 Aalto University Deptartment of Computer Science and Engineering [email protected] Abstract. We present a method to speed up approximate string match- ing by mapping the factual alphabet to a smaller alphabet. We apply the alphabet reduction scheme to a tuned version of the approximate Boyer– Moore algorithm utilizing the Four-Russians technique. Our experiments show that the alphabet reduction makes the algorithm faster. Especially in the k-mismatch case, the new variation is faster than earlier algorithms for English data with small values of k. 1 Introduction The approximate string matching problem is defined as follows. We have a pat- tern P [1...m]ofm characters drawn from an alphabet Σ of size σ,atextT [1...n] of n characters over the same alphabet, and an integer k. We need to find all such positions i of the text that the distance between the pattern and a sub- string of the text ending at that position is at most k.Inthek-difference problem the distance between two strings is the standard edit distance where substitu- tions, deletions, and insertions are allowed. The k-mismatch problem is a more restricted one using the Hamming distance where only substitutions are allowed. Among the most cited papers on approximate string matching are the classical articles [1,2] by Esko Ukkonen. Besides them he has studied this topic extensively [3,4,5,6,7,8,9,10,11].
    [Show full text]
  • Searching All Seeds of Strings with Hamming Distance Using Finite Automata
    Proceedings of the International MultiConference of Engineers and Computer Scientists 2009 Vol I IMECS 2009, March 18 - 20, 2009, Hong Kong Searching All Seeds of Strings with Hamming Distance using Finite Automata Ondřej Guth, Bořivoj Melichar ∗ Abstract—Seed is a type of a regularity of strings. Finite automata provide common formalism for many al- A restricted approximate seed w of string T is a fac- gorithms in the area of text processing (stringology), in- tor of T such that w covers a superstring of T under volving forward exact and approximate pattern match- some distance rule. In this paper, the problem of all ing and searching for borders, periods, and repetitions restricted seeds with the smallest Hamming distance [7], backward pattern matching [8], pattern matching in is studied and a polynomial time and space algorithm a compressed text [9], the longest common subsequence for solving the problem is presented. It searches for [10], exact and approximate 2D pattern matching [11], all restricted approximate seeds of a string with given limited approximation using Hamming distance and and already mentioned computing approximate covers it computes the smallest distance for each found seed. [5, 6] and exact covers [12] and seeds [2] in generalized The solution is based on a finite (suffix) automata ap- strings. Therefore, we would like to further extend the set proach that provides a straightforward way to design of problems solved using finite automata. Such a problem algorithms to many problems in stringology. There- is studied in this paper. fore, it is shown that the set of problems solvable using finite automata includes the one studied in this Finite automaton as a data structure may be easily imple- paper.
    [Show full text]
  • Approximate Boyer-Moore String Matching
    APPROXIMATE BOYER-MOORE STRING MATCHING JORMA TARHIO AND ESKO UKKONEN University of Helsinki, Department of Computer Science Teollisuuskatu 23, SF-00510 Helsinki, Finland Draft Abstract. The Boyer-Moore idea applied in exact string matching is generalized to approximate string matching. Two versions of the problem are considered. The k mismatches problem is to find all approximate occurrences of a pattern string (length m) in a text string (length n) with at most k mis- matches. Our generalized Boyer-Moore algorithm is shown (under a mild 1 independence assumption) to solve the problem in expected time O(kn( + m – k k)) where c is the size of the alphabet. A related algorithm is developed for the c k differences problem where the task is to find all approximate occurrences of a pattern in a text with ≤ k differences (insertions, deletions, changes). Experimental evaluation of the algorithms is reported showing that the new algorithms are often significantly faster than the old ones. Both algorithms are functionally equivalent with the Horspool version of the Boyer-Moore algorithm when k = 0. Key words: String matching, edit distance, Boyer-Moore algorithm, k mismatches problem, k differences problem AMS (MOS) subject classifications: 68C05, 68C25, 68H05 Abbreviated title: Approximate Boyer-Moore Matching 2 1. Introduction The fastest known exact string matching algorithms are based on the Boyer- Moore idea [BoM77, KMP77]. Such algorithms are “sublinear” on the average in the sense that it is not necessary to check every symbol in the text. The larger is the alphabet and the longer is the pattern, the faster the algorithm works.
    [Show full text]
  • B-Bit Sketch Trie: Scalable Similarity Search on Integer Sketches
    b-Bit Sketch Trie: Scalable Similarity Search on Integer Sketches Shunsuke Kanda Yasuo Tabei RIKEN Center for Advanced Intelligence Project RIKEN Center for Advanced Intelligence Project Tokyo, Japan Tokyo, Japan [email protected] [email protected] Abstract—Recently, randomly mapping vectorial data to algorithms intending to build sketches of non-negative inte- strings of discrete symbols (i.e., sketches) for fast and space- gers (i.e., b-bit sketches) have been proposed for efficiently efficient similarity searches has become popular. Such random approximating various similarity measures. Examples are b-bit mapping is called similarity-preserving hashing and approximates a similarity metric by using the Hamming distance. Although minwise hashing (minhash) [12]–[14] for Jaccard similarity, many efficient similarity searches have been proposed, most of 0-bit consistent weighted sampling (CWS) for min-max ker- them are designed for binary sketches. Similarity searches on nel [15], and 0-bit CWS for generalized min-max kernel [16]. integer sketches are in their infancy. In this paper, we present Thus, developing scalable similarity search methods for b-bit a novel space-efficient trie named b-bit sketch trie on integer sketches is a key issue in large-scale applications of similarity sketches for scalable similarity searches by leveraging the idea behind succinct data structures (i.e., space-efficient data structures search. while supporting various data operations in the compressed Similarity searches on binary sketches are classified
    [Show full text]
  • Problem Set 7 Solutions
    Introduction to Algorithms November 18, 2005 Massachusetts Institute of Technology 6.046J/18.410J Professors Erik D. Demaine and Charles E. Leiserson Handout 25 Problem Set 7 Solutions Problem 7-1. Edit distance In this problem you will write a program to compute edit distance. This problem is mandatory. Failure to turn in a solution will result in a serious and negative impact on your term grade! We advise you to start this programming assignment as soon as possible, because getting all the details right in a program can take longer than you think. Many word processors and keyword search engines have a spelling correction feature. If you type in a misspelled word x, the word processor or search engine can suggest a correction y. The correction y should be a word that is close to x. One way to measure the similarity in spelling between two text strings is by “edit distance.” The notion of edit distance is useful in other fields as well. For example, biologists use edit distance to characterize the similarity of DNA or protein sequences. The edit distance d(x; y) of two strings of text, x[1 : : m] and y[1 : : n], is defined to be the minimum possible cost of a sequence of “transformation operations” (defined below) that transforms string x[1 : : m] into string y[1 : : n].1 To define the effect of the transformation operations, we use an auxiliary string z[1 : : s] that holds the intermediate results. At the beginning of the transformation sequence, s = m and z[1 : : s] = x[1 : : m] (i.e., we start with string x[1 : : m]).
    [Show full text]
  • Large Scale Hamming Distance Query Processing
    Large Scale Hamming Distance Query Processing Alex X. Liu, Ke Shen, Eric Torng Department of Computer Science and Engineering Michigan State University East Lansing, MI 48824, U.S.A. {alexliu, shenke, torng}@cse.msu.edu Abstract—Hamming distance has been widely used in many a new solution to the Hamming distance range query problem. application domains, such as near-duplicate detection and pattern The basic idea is to hash each document into a 64-bit binary recognition. We study Hamming distance range query problems, string, called a fingerprint or a sketch, using the simhash where the goal is to find all strings in a database that are within a Hamming distance bound k from a query string. If k is fixed, algorithm in [11]. Simhash has the following property: if two we have a static Hamming distance range query problem. If k is documents are near-duplicates in terms of content, then the part of the input, we have a dynamic Hamming distance range Hamming distance between their fingerprints will be small. query problem. For the static problem, the prior art uses lots of Similarly, Miller et al. proposed a scheme for finding a song memory due to its aggressive replication of the database. For the that includes a given audio clip [20], [21]. They proposed dynamic range query problem, as far as we know, there is no space and time efficient solution for arbitrary databases. In this converting the audio clip to a binary string (and doing the paper, we first propose a static Hamming distance range query same for all the songs in the given database) and then finding algorithm called HEngines, which addresses the space issue in the nearest neighbors to the given binary string in the database prior art by dynamically expanding the query on the fly.
    [Show full text]
  • Parity Check Code Repetition Code 10010011 11100001
    Lecture Notes Charles Cusack Communication Errors Introduction Implementations of memory, storage, and communication are not perfect. Therefore, it is possible that when I store a number on a hard drive or send it over a network, a slightly different number is actually stored/received. There are techniques to deal with this. Definition An error detection (correction) code is a method of encoding data in so that if a certain number of errors has occurred during transmission, it can be detected (corrected). Definition The hamming distance between two bit strings is the number of bits where they differ. Example The hamming distance between 10010011 and 11100001 is 4 since 1001 00 11 these bit patterns differ in bits 2, 3, 4, and 7. 1110 00 01 Parity Check Code The simplest method of detecting errors is by using a parity bit . To use a parity bit, one simply appends one additional bit (to the beginning or end—it doesn’t matter which as long as you are consistent) to a bit string so that the number of 1s in the string is odd (or even). To detect an error you add the number of ‘1’ bits. If it is not odd (even) you know an error has occurred. Example The bit string 10010011 is encoded as 100100111 using an odd parity check code, and the bit string 11110001 is encoded as 111100010 using an odd parity check code. Question • How many errors can this method detect? What happens if more errors occur? • How many errors can this method correct? Repetition Code A repetition code encodes a bit string by appending k copies of the bit string.
    [Show full text]
  • Bit-Parallel Approximate String Matching Under Hamming Distance
    Aalto University School of Science Degree Programme in Computer Science and Engineering Tommi Hirvola Bit-parallel approximate string matching under Hamming distance Master's Thesis Espoo, July 21, 2016 Supervisors: Professor Jorma Tarhio Advisor: Professor Jorma Tarhio Aalto University School of Science ABSTRACT OF Degree Programme in Computer Science and Engineering MASTER'S THESIS Author: Tommi Hirvola Title: Bit-parallel approximate string matching under Hamming distance Date: July 21, 2016 Pages: 87 Major: Software engineering Code: T-106 Supervisors: Professor Jorma Tarhio Advisor: Professor Jorma Tarhio In many situations, an exact search of a pattern in a text does not find all in- teresting parts of the text. For example, exact matching can be rendered useless by spelling mistakes or other errors in the text or pattern. Approximate string matching addresses the problem by allowing a certain amount of error in occur- rences of the pattern in the text. This thesis considers the approximate matching problem known as string matching with k mismatches, in which k characters are allowed to mismatch. The main contribution of the thesis consists of new algorithms for approximate string matching with k mismatches based on the bit-parallelism technique. Ex- periments show that the proposed algorithms are competitive with the state-of- the-art algorithms in practice. A significant advantage of the novel algorithms is algorithmic simplicity and flexibility, which is demonstrated by deriving efficient algorithms to several problems involving approximate string matching under the k-mismatches model. A bit-parallel algorithm is presented for online approximate matching of circular strings with k mismatches, which is the first average-optimal solution that is not based on filtering.
    [Show full text]
  • 3. Approximate String Matching
    3. Approximate String Matching Often in applications we want to search a text for something that is similar to the pattern but not necessarily exactly the same. To formalize this problem, we have to specify what does “similar” mean. This can be done by defining a similarity or a distance measure. A natural and popular distance measure for strings is the edit distance, also known as the Levenshtein distance. 109 Edit distance The edit distance ed(A, B) of two strings A and B is the minimum number of edit operations needed to change A into B. The allowed edit operations are: S Substitution of a single character with another character. I Insertion of a single character. D Deletion of a single character. Example 3.1: Let A = Lewensteinn and B = Levenshtein. Then ed(A, B) = 3. The set of edit operations can be described with an edit sequence: NNSNNNINNNND or with an alignment: Lewens-teinn Levenshtein- In the edit sequence, N means No edit. 110 There are many variations and extension of the edit distance, for example: Hamming distance allows only the subtitution operation. • Damerau–Levenshtein distance adds an edit operation: • T Transposition swaps two adjacent characters. With weighted edit distance, each operation has a cost or weight, • which can be other than one. Allow insertions and deletions (indels) of factors at a cost that is lower • than the sum of character indels. We will focus on the basic Levenshtein distance. Levenshtein distance has the following two useful properties, which are not shared by all variations (exercise): Levenshtein distance is a metric.
    [Show full text]