Lecture 4: Hashing

Total Page:16

File Type:pdf, Size:1020Kb

Lecture 4: Hashing Lecture 4: Hashing Rafael Oliveira University of Waterloo Cheriton School of Computer Science [email protected] September 21, 2020 1 / 78 Overview Introduction Hash Functions Why is hashing? How to hash? Succinctness of Hash Functions Coping with randomness Universal Hashing Hashing using 2-universal families Perfect Hashing Acknowledgements 2 / 78 Computational Model Before we talk about hash functions, we need to state our model of computation: Definition (Word RAM model) In the word RAMa model: all elements are integers that fit in a machine word of w bits Basic operations (comparison, arithmetic, bitwise) on such words take Θ(1) time We can also access any position in the array in Θ(1) time aRAM stands for Random Access Model 3 / 78 Naive approach: use an array A of m elements, initially A[i] = 0 for all i, and when a key is inserted, set A[i] = 1. Insertion: O(1), Deletion: O(1), Search: O(1) Memory: O(m log(m)) (this is very bad!) Want to also achieve optimal memory O(n log(m)). For this we will use a technique called hashing. A hash function is a function h : U ! [0; n − 1], where jUj = m >> n. A hash table is a data structure that consists of: a table T with n cells [0; n − 1], each cell storing O(log(m)) bits a hash function h : U ! [0; n − 1] From now on, we will define memory as # of cells. What is hashing? We want to store n elements (keys) from the set U = f0; 1;:::; m − 1g, where m >> n, in a data structure that supports insertions, deletions, search \as efficiently as possible." 4 / 78 Insertion: O(1), Deletion: O(1), Search: O(1) Memory: O(m log(m)) (this is very bad!) Want to also achieve optimal memory O(n log(m)). For this we will use a technique called hashing. A hash function is a function h : U ! [0; n − 1], where jUj = m >> n. A hash table is a data structure that consists of: a table T with n cells [0; n − 1], each cell storing O(log(m)) bits a hash function h : U ! [0; n − 1] From now on, we will define memory as # of cells. What is hashing? We want to store n elements (keys) from the set U = f0; 1;:::; m − 1g, where m >> n, in a data structure that supports insertions, deletions, search \as efficiently as possible." Naive approach: use an array A of m elements, initially A[i] = 0 for all i, and when a key is inserted, set A[i] = 1. 5 / 78 Want to also achieve optimal memory O(n log(m)). For this we will use a technique called hashing. A hash function is a function h : U ! [0; n − 1], where jUj = m >> n. A hash table is a data structure that consists of: a table T with n cells [0; n − 1], each cell storing O(log(m)) bits a hash function h : U ! [0; n − 1] From now on, we will define memory as # of cells. What is hashing? We want to store n elements (keys) from the set U = f0; 1;:::; m − 1g, where m >> n, in a data structure that supports insertions, deletions, search \as efficiently as possible." Naive approach: use an array A of m elements, initially A[i] = 0 for all i, and when a key is inserted, set A[i] = 1. Insertion: O(1), Deletion: O(1), Search: O(1) Memory: O(m log(m)) (this is very bad!) 6 / 78 What is hashing? We want to store n elements (keys) from the set U = f0; 1;:::; m − 1g, where m >> n, in a data structure that supports insertions, deletions, search \as efficiently as possible." Naive approach: use an array A of m elements, initially A[i] = 0 for all i, and when a key is inserted, set A[i] = 1. Insertion: O(1), Deletion: O(1), Search: O(1) Memory: O(m log(m)) (this is very bad!) Want to also achieve optimal memory O(n log(m)). For this we will use a technique called hashing. A hash function is a function h : U ! [0; n − 1], where jUj = m >> n. A hash table is a data structure that consists of: a table T with n cells [0; n − 1], each cell storing O(log(m)) bits a hash function h : U ! [0; n − 1] From now on, we will define memory as # of cells. 7 / 78 Why is hashing useful? Designing efficient data structures (dictionaries) for searching Data streaming algorithms Derandomization Cryptography Complexity Theory many more 8 / 78 Will settle for: # collisions small with high probability. Ideally, want hash function to map different keys into different locations. Definition (Collision) We say that a collision happens for hash function h with inputs x; y 2 U if x 6= y and h(x) = h(y). By pigeonhole principle, impossible to achieve without knowing keys in advance. Challenges in Hashing Setup: Universe U = f0;:::; m − 1g of size m >> n where n is the size of the range of our hash function h : U ! [0; n − 1] Store O(n) elements of U (keys) in hash table T (which has n cells) 9 / 78 Will settle for: # collisions small with high probability. Challenges in Hashing Setup: Universe U = f0;:::; m − 1g of size m >> n where n is the size of the range of our hash function h : U ! [0; n − 1] Store O(n) elements of U (keys) in hash table T (which has n cells) Ideally, want hash function to map different keys into different locations. Definition (Collision) We say that a collision happens for hash function h with inputs x; y 2 U if x 6= y and h(x) = h(y). By pigeonhole principle, impossible to achieve without knowing keys in advance. 10 / 78 Challenges in Hashing Setup: Universe U = f0;:::; m − 1g of size m >> n where n is the size of the range of our hash function h : U ! [0; n − 1] Store O(n) elements of U (keys) in hash table T (which has n cells) Ideally, want hash function to map different keys into different locations. Definition (Collision) We say that a collision happens for hash function h with inputs x; y 2 U if x 6= y and h(x) = h(y). By pigeonhole principle, impossible to achieve without knowing keys in advance. Will settle for: # collisions small with high probability. 11 / 78 Simplest version to keep in mind: 1 Pr [h(x) = h(y)] ≤ 8x 6= y 2 U h2R H poly(n) Assumptions: keys are independent from hash function we choose. we do not know keys in advance (even if we did, nontrivial problem!) Question Still could have collisions. How do we handle them? Our solution: family of hash functions Construct family of hash functions H such that the number of collisions is small with high probability, when we pick hash function uniformly at random from the family H. 12 / 78 Assumptions: keys are independent from hash function we choose. we do not know keys in advance (even if we did, nontrivial problem!) Question Still could have collisions. How do we handle them? Our solution: family of hash functions Construct family of hash functions H such that the number of collisions is small with high probability, when we pick hash function uniformly at random from the family H. Simplest version to keep in mind: 1 Pr [h(x) = h(y)] ≤ 8x 6= y 2 U h2R H poly(n) 13 / 78 Our solution: family of hash functions Construct family of hash functions H such that the number of collisions is small with high probability, when we pick hash function uniformly at random from the family H. Simplest version to keep in mind: 1 Pr [h(x) = h(y)] ≤ 8x 6= y 2 U h2R H poly(n) Assumptions: keys are independent from hash function we choose. we do not know keys in advance (even if we did, nontrivial problem!) Question Still could have collisions. How do we handle them? 14 / 78 This setting is same as our balls-and-bins setting! So, if we have to store n keys: Expected number of keys in a location:1 maximum number of collisions (max load) in one particular location: O(log n= log log n) keys Solving collisions: store all keys hashed into location i by a linked list. Known as chain hashing. Could also pick two random hash functions and use power of two choices. Collision bound becomes O(log log n). Random Hash Functions? Natural to consider following approach: From all functions h : U ! [0; n − 1], just pick one uniformly at random. 15 / 78 So, if we have to store n keys: Expected number of keys in a location:1 maximum number of collisions (max load) in one particular location: O(log n= log log n) keys Solving collisions: store all keys hashed into location i by a linked list. Known as chain hashing. Could also pick two random hash functions and use power of two choices. Collision bound becomes O(log log n). Random Hash Functions? Natural to consider following approach: From all functions h : U ! [0; n − 1], just pick one uniformly at random. This setting is same as our balls-and-bins setting! 16 / 78 Solving collisions: store all keys hashed into location i by a linked list. Known as chain hashing. Could also pick two random hash functions and use power of two choices. Collision bound becomes O(log log n).
Recommended publications
  • Key-Recovery Attacks on Universal Hash Function Based MAC Algorithms
    Key-Recovery Attacks on Universal Hash Function Based MAC Algorithms Helena Handschuh1 and Bart Preneel2,3 1 Spansion, 105 rue Anatole France 92684 Levallois-Perret Cedex, France [email protected] 2 Katholieke Universiteit Leuven, Dept. Electrical Engineering-ESAT/COSIC, Kasteelpark Arenberg 10, bus 2446, B-3001 Leuven, Belgium [email protected] 3 IBBT, Van Crommenlaan, B-9000 Gent Abstract. This paper discusses key recovery and universal forgery at- tacks on several MAC algorithms based on universal hash functions. The attacks use a substantial number of verification queries but eventually allow for universal forgeries instead of existential or multiple forgeries. This means that the security of the algorithms completely collapses once a few forgeries are found. Some of these attacks start off by exploiting a weak key property, but turn out to become full-fledged divide and conquer attacks because of the specific structure of the universal hash functions considered. Partial information on a secret key can be exploited too, in the sense that it renders some key recovery attacks practical as soon as a few key bits are known. These results show that while universal hash functions offer provable security, high speeds and parallelism, their simple combinatorial properties make them less robust than conventional message authentication primitives. 1 Introduction Message Authentication Code (MAC) algorithms are symmetric cryptographic primitives that allow senders and receivers who share a common secret key to make sure the contents of a transmitted message has not been tampered with. Three main types of constructions for MAC algorithms can be found in the literature: constructions based on block ciphers, based on hash functions and based on universal hash functions.
    [Show full text]
  • CS 473: Algorithms, Fall 2019
    CS 473: Algorithms, Fall 2019 Universal and Perfect Hashing Lecture 10 September 26, 2019 Chandra and Michael (UIUC) cs473 1 Fall 2019 1 / 45 Today's lecture: Review pairwise independence and related constructions (Strongly) Universal hashing Perfect hashing Announcements and Overview Pset 4 released and due on Thursday, October 3 at 10am. Note one day extension over usual deadline. Midterm 1 is on Monday, Oct 7th from 7-9.30pm. More details and conflict exam information will be posted on Piazza. Next pset will be released after the midterm exam. Chandra and Michael (UIUC) cs473 2 Fall 2019 2 / 45 Announcements and Overview Pset 4 released and due on Thursday, October 3 at 10am. Note one day extension over usual deadline. Midterm 1 is on Monday, Oct 7th from 7-9.30pm. More details and conflict exam information will be posted on Piazza. Next pset will be released after the midterm exam. Today's lecture: Review pairwise independence and related constructions (Strongly) Universal hashing Perfect hashing Chandra and Michael (UIUC) cs473 2 Fall 2019 2 / 45 Part I Review Chandra and Michael (UIUC) cs473 3 Fall 2019 3 / 45 Pairwise independent random variables Definition Random variables X1; X2;:::; Xn from a range B are pairwise independent if for all 1 ≤ i < j ≤ n and for all b; b0 2 B, 0 0 Pr[Xi = b; Xj = b ] = Pr[Xi = b] · Pr[Xj = b ] : Chandra and Michael (UIUC) cs473 4 Fall 2019 4 / 45 Interesting case: n = m = p where p is a prime number Pick a; b uniformly at random from f0; 1; 2;:::; p − 1g Set Xi = ai + b Only need to store a; b.
    [Show full text]
  • Security of Authentication Based on a Universal Hash-Function Family
    c de Gruyter 2010 J. Math. Crypt. 4 (2010), 1–24 DOI 10.1515 / JMC.2010.xxx The power of primes: security of authentication based on a universal hash-function family Basel Alomair, Andrew Clark and Radha Poovendran Communicated by xxx Abstract. Message authentication codes (MACs) based on universal hash-function families are be- coming increasingly popular due to their fast implementation. In this paper, we investigate a family of universal hash functions that has been appeared repeatedly in the literature and provide a detailed algebraic analysis for the security of authentication codes based on this universal hash family. In particular, the universal hash family under analysis, as appeared in the literature, uses operation in the finite field Zp. No previous work has studied the extension of such universal hash family when computations are performed modulo a non-prime integer n. In this work, we provide the first such analysis. We investigate the security of authentication when computations are performed over arbi- trary finite integer rings Zn and derive an explicit relation between the prime factorization of n and the bound on the probability of successful forgery. More specifically, we show that the probability of successful forgery against authentication codes based on such a universal hash-function family is bounded by the reciprocal of the smallest prime factor of the modulus n. Keywords. Cryptography, authentication, finite integer rings, universal hash-function families. AMS classification. 11T71, 94A60, 94A62. 1 Introduction and related work Message authentication code (MAC) algorithms can be categorized, based on their se- curity, into unconditionally and conditionally secure MACs.
    [Show full text]
  • New Bounds for Almost Universal Hash Functions
    New bounds for almost universal hash functions Abstract Using combinatorial analysis and the pigeon-hole principle, we introduce a new lower (or combinatorial) bound for the key length in an almost (both pairwise and l-wise) universal hash function. The advantage of this bound is that the key length only grows in proportion to the logarithm of the message length as the collision probability moves away from its theoretical minimum by an arbitrarily small and positive value. Conventionally bounds for various families of universal hash functions have been derived by using the equivalence between pairwise universal hash functions and error-correcting codes, and indeed in this paper another (very similar) bound for almost universal hash functions can be calculated from the Singleton bound in coding theory. However, the latter, perhaps unexpectedly, will be shown to be not as tight as our combinatorial one. To the best of our knowledge, this is the first time that combinatorial analysis has been demonstrated to yield a better universal hash function bound than the use of the equivalence, and we will explain why there is such a mismatch. This work therefore potentially opens the way to re-examining many bounds for a number of families of universal hash functions, which have been solely derived from bounds in coding theory and/or other combinatorial objects. 1 Introduction and contribution Universal hash function H with parameters (, K, M, b) was introduced by Carter and Wegman [9, 43]. Each family, which is indexed by a K-bit key k, consists of 2K hash functions mapping a message representable by M bits into a b-bit hash output.
    [Show full text]
  • High Speed Hashing for Integers and Strings
    High Speed Hashing for Integers and Strings Mikkel Thorup May 12, 2020 Abstract These notes describe the most efficient hash functions currently known for hashing integers and strings. These modern hash functions are often an order of magnitude faster than those presented in standard text books. They are also simpler to implement, and hence a clear win in practice, but their analysis is harder. Some of the most practical hash functions have only appeared in theory papers, and some of them require combining results from different theory papers. The goal here is to combine the information in lecture-style notes that can be used by theoreticians and practitioners alike, thus making these practical fruits of theory more widely accessible. 1 Hash functions The concept of truly independent hash functions is extremely useful in the design of randomized algorithms. We have a large universe U of keys, e.g., 64-bit numbers, that we wish to map ran- domly to a range [m]= 0,...,m 1 of hash values. A truly random hash function h : U [m] { − } → assigns an independent uniformly random variable h(x) to each key in x. The function h is thus a U -dimensional random variable, picked uniformly at random among all functions from U to [m]. | | Unfortunately truly random hash functions are idealized objects that cannot be implemented. More precisely, to represent a truly random hash function, we need to store at least U log2 m bits, and in most applications of hash functions, the whole point in hashing is that the universe| | is much arXiv:1504.06804v9 [cs.DS] 9 May 2020 too large for such a representation (at least not in fast internal memory).
    [Show full text]
  • Universal and Perfect Hashing
    Lecture 10 Universal and Perfect Hashing 10.1 Overview Hashing is a great practical tool, with an interesting and subtle theory too. In addition to its use as a dictionary data structure, hashing also comes up in many different areas, including cryptography and complexity theory. In this lecture we describe two important notions: universal hashing (also known as universal hash function families) and perfect hashing. Material covered in this lecture includes: • The formal setting and general idea of hashing. • Universal hashing. • Perfect hashing. 10.2 Introduction We will be looking at the basic dictionary problem we have been discussing so far and will consider two versions, static and dynamic: • Static: Given a set S of items, we want to store them so that we can do lookups quickly. E.g., a fixed dictionary. • Dynamic: here we have a sequence of insert, lookup, and perhaps delete requests. We want to do these all efficiently. For the first problem we could use a sorted array with binary search for lookups. For the second we could use a balanced search tree. However, hashing gives an alternative approach that is often the fastest and most convenient way to solve these problems. For example, suppose you are writing an AI-search program, and you want to store situations that you’ve already solved (board positions or elements of state-space) so that you don’t redo the same computation when you encounter them again. Hashing provides a simple way of storing such information. There are also many other uses in cryptography, networks, complexity theory. 52 10.3.
    [Show full text]
  • Regular and Almost Universal Hashing: an Efficient Implementation
    Regular and almost universal hashing: an efficient implementation D. Ivanchykhin1, S. Ignatchenko1, D. Lemire2 1OLogN Technologies AG, Triesen, Liechtenstein 2LICEF Research Center, TELUQ, Montreal, QC, Canada SUMMARY Random hashing can provide guarantees regarding the performance of data structures such as hash tables— even in an adversarial setting. Many existing families of hash functions are universal: given two data objects, the probability that they have the same hash value is low given that we pick hash functions at random. However, universality fails to ensure that all hash functions are well behaved. We might further require regularity: when picking data objects at random they should have a low probability of having the same hash value, for any fixed hash function. We present the efficient implementation of a family of non-cryptographic hash functions (PM+) offering good running times, good memory usage as well as distinguishing theoretical guarantees: almost universality and component-wise regularity. On a variety of platforms, our implementations are comparable to the state of the art in performance. On recent Intel processors, PM+ achieves a speed of 4.7 bytes per cycle for 32-bit outputs and 3.3 bytes per cycle for 64-bit outputs. We review vectorization through SIMD instructions (e.g., AVX2) and optimizations for superscalar execution. KEY WORDS: performance; measurement; random hashing, universal hashing, non-cryptographic hashing, avalanche effect 1. INTRODUCTION Hashing is ubiquitous in software. For example, most programming languages support hash tables, either directly, or via libraries. However, while many computer science textbooks consider random hashing, most software libraries use deterministic (i.e., non-random) hashing.
    [Show full text]
  • Hashing Techniques: a Survey and Taxonomy
    11 Hashing Techniques: A Survey and Taxonomy LIANHUA CHI, IBM Research, Melbourne, Australia XINGQUAN ZHU, Florida Atlantic University, Boca Raton, FL; Fudan University, Shanghai, China With the rapid development of information storage and networking technologies, quintillion bytes of data are generated every day from social networks, business transactions, sensors, and many other domains. The increasing data volumes impose significant challenges to traditional data analysis tools in storing, processing, and analyzing these extremely large-scale data. For decades, hashing has been one of the most effective tools commonly used to compress data for fast access and analysis, as well as information integrity verification. Hashing techniques have also evolved from simple randomization approaches to advanced adaptive methods considering locality, structure, label information, and data security, for effective hashing. This survey reviews and categorizes existing hashing techniques as a taxonomy, in order to provide a comprehensive view of mainstream hashing techniques for different types of data and applications. The taxonomy also studies the uniqueness of each method and therefore can serve as technique references in understanding the niche of different hashing mechanisms for future development. Categories and Subject Descriptors: A.1 [Introduction and Survey]: Hashing General Terms: Design, algorithms Additional Key Words and Phrases: Hashing, compression, dimension reduction, data coding, cryptographic hashing ACM Reference Format: Lianhua Chi and Xingquan Zhu. 2017. Hashing techniques: A survey and taxonomy. ACM Comput. Surv. 50, 1, Article 11 (April 2017), 36 pages. DOI: http://dx.doi.org/10.1145/3047307 1. INTRODUCTION Recent advancement in information systems, including storage devices and networking techniques, have resulted in many applications generating large volumes of data and requiring large storage, fast delivery, and quick analysis.
    [Show full text]
  • A Fast Single-Key Two-Level Universal Hash Function
    A Fast Single-Key Two-Level Universal Hash Function Debrup Chakraborty, Sebati Ghosh and Palash Sarkar Indian Statistical Institute 203, B.T.Road, Kolkata, India - 700108. fdebrup,sebati r,[email protected] November 23, 2016 Abstract Universal hash functions based on univariate polynomials are well known, e.g. Poly1305 and GHASH. Using Horner's rule to evaluate such hash functions require ` − 1 field multiplications for hashing a message consisting of ` blocks where each block is one field element. A faster method is based on the class of Bernstein-Rabin-Winograd (BRW) polynomials which require b`=2c multipli- cations and blg `c squarings for ` ≥ 3 blocks. Though this is significantly smaller than Horner's rule based hashing, implementation of BRW polynomials for variable length messages present significant difficulties. In this work, we propose a two-level hash function where BRW polynomial based hashing is done at the lower level and Horner's rule based hashing is done at the higher level. The BRW polynomial based hashing is applied to a fixed number of blocks and hence the difficulties in handling variable length messages is avoided. Even though the hash function has two levels, we show that it is sufficient to use a single field element as the hash key. The basic idea is instantiated to propose two new hash functions, one which hashes a single binary string and the other can hash a vector of binary strings. We describe two actual implementations, one over F2128 and the other over F2256 both using the pclmulqdq instruction available in modern Intel processors.
    [Show full text]
  • Lightweight Macs from Universal Hash Functions Sébastien Duval, Gaëtan Leurent
    Lightweight MACs from Universal Hash Functions Sébastien Duval, Gaëtan Leurent To cite this version: Sébastien Duval, Gaëtan Leurent. Lightweight MACs from Universal Hash Functions. CARDIS 2019 - 18th Smart Card Research and Advanced Application Conference, Nov 2019, Prague, Czech Republic. hal-02424904 HAL Id: hal-02424904 https://hal.inria.fr/hal-02424904 Submitted on 28 Dec 2019 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Lightweight MACs from Universal Hash Functions S´ebastienDuval1;2;3 and Ga¨etanLeurent1 1 Inria, Paris, France 2 Sorbonne Universit´es,Paris, France 3 UCLouvain, Louvain-la-Neuve, Belgium Abstract. Lightweight cryptography is a topic of growing importance, with the goal to secure the communication of low-end devices that are not powerful enough to use conventional cryptography. There have been many recent proposals of lightweight block ciphers, but comparatively few results on lightweight Message Authentication Codes (MACs). Therefore, this paper focuses on lightweight MACs. We review some ex- isting constructions, and revisit the choices made in mainstream MACs with a focus on lightweight cryptography. We consider MACs based on universal hash functions, because they offer information theoretic secu- rity, can be implemented efficiently and are widely used in conventional cryptography.
    [Show full text]
  • 12 Hash Tables
    Algorithms Lecture 12: Hash Tables [Fa’10] Calvin: There! I finished our secret code! Hobbes: Let’s see. Calvin: I assigned each letter a totally random number, so the code will be hard to crack. For letter “A”, you write 3,004,572,688. “B” is 28,731,569½. Hobbes: That’s a good code all right. Calvin: Now we just commit this to memory. Calvin: Did you finish your map of our neighborhood? Hoobes: Not yet. How many bricks does the front walk have? — Bill Watterson, “Calvin and Hobbes” (August 23, 1990) 12 Hash Tables 12.1 Introduction A hash table is a data structure for storing a set of items, so that we can quickly determine whether an item is or is not in the set. The basic idea is to pick a hash function h that maps every possible item x to a small integer h(x). Then we store x in slot h(x) in an array. The array is the hash table. Let’s be a little more specific. We want to store a set of n items. Each item is an element of some finite1 set U called the universe; we use u to denote the size of the universe, which is just the number of items in U. A hash table is an array T[1 .. m], where m is another positive integer, which we call the table size. Typically, m is much smaller than u.A hash function is any function of the form h: U 0, 1, . , m 1 , ! f − g mapping each possible item in U to a slot in the hash table.
    [Show full text]
  • UMAC: Fast and Secure Message Authentication
    The proceedings version of this paper appears in Advances in Cryptology—CRYPTO ’99 [7]. This is the full version. It is available from www.cs.ucdavis.edu/~rogaway/umac/ UMAC: Fast and Secure Message Authentication J. Black ∗ S. Halevi † H. Krawczyk ‡ T. Krovetz ∗ P. Rogaway ∗ August 4, 1999 Abstract We describe a message authentication algorithm, UMAC, which authenticates messages (in software, on contemporary machines) roughly an order of magnitude faster than current practice (e.g., HMAC-SHA1), and about twice as fast as previously reported times for the universal hash- function family MMH. To achieve such speeds, UMAC uses a new universal hash-function family, NH, and a design which allows effective exploitation of SIMD parallelism. The “cryptographic” work of UMAC is done using standard primitives of the user’s choice, such as a block cipher or cryptographic hash function; no new heuristic primitives are developed here. Instead, the security of UMAC is rigorously proven, in the sense of giving exact and quantitatively strong results which demonstrate an inability to forge UMAC-authenticated messages assuming an inability to break the underlying cryptographic primitive. Unlike conventional, inherently serial MACs, UMAC is parallelizable, and will have ever-faster implementation speeds as machines offer up increasing amounts of parallelism. We envision UMAC as a practical algorithm for next-generation message authentication. Key words: Fast software cryptography, Message authentication codes, Provable security, Universal hashing. ∗ Dept. of Computer Science, University of California, Davis, California, 95616, USA. E-mail: {blackj, krovetz, rogaway}@cs.ucdavis.edu WWW: www.cs.ucdavis.edu/~{blackj, krovetz, rogaway}/ † IBM T.J.
    [Show full text]