Provably Secure Steganography

Total Page:16

File Type:pdf, Size:1020Kb

Provably Secure Steganography IEEE TRANSACTIONS ON COMPUTERS, VOL. 58, NO. X, XXX 2009 1 Provably Secure Steganography Nicholas Hopper, Luis von Ahn, and John Langford Abstract—Steganography is the problem of hiding secret messages in “innocent-looking” public communication so that the presence of the secret messages cannot be detected. This paper introduces a cryptographic formalization of steganographic security in terms of computational indistinguishability from a channel, an indexed family of probability distributions on cover messages. We use cryptographic and complexity-theoretic proof techniques to show that the existence of one-way functions and the ability to sample from the channel are necessary conditions for secure steganography. We then construct a steganographic protocol, based on rejection sampling from the channel, that is provably secure and has nearly optimal bandwidth under these conditions. This is the first known example of a general provably secure steganographic protocol. We also give the first formalization of “robust” steganography, where an adversary attempts to remove any hidden messages without unduly disrupting the cover channel. We give a necessary condition on the amount of disruption the adversary is allowed in terms of a worst case measure of mutual information. We give a construction that is provably secure and computationally efficient and has nearly optimal bandwidth, assuming repeatable access to the channel distribution. Index Terms—Steganography, covert channels, provable security. Ç 1INTRODUCTION classic problem of computer security is the mitigation Steganographic “protocols” have a long and intriguing Aof covert channels. First introduced by Lampson [1], a history that predates covert channels, stretching back to covert channel in a (single-host or distributed) computer antiquity. For example, Kahn [3] relates stories of World system can be roughly defined as any means by which two War II prisoners, spies, and soldiers sending secret processes or users can exchange information in violation of messages—written in invisible ink, hidden in love letters security policy. While the exact detection of usable covert (e.g., hiding the message in the first character of each channels in a system is undecidable, many conservative sentence), or using other “physical” steganographic approaches exist to detect and eliminate all potential covert techniques—to circumvent inspections by both the Allied channels; for example, the US Department of Defense and Axis governments. In response, postal censors crossed “Light Pink Book” [2] on covert channel analysis includes out anything that looked like sensitive information (e.g., detailed procedures to find and eliminate covert channels. long strings of digits) and they prosecuted individuals Unfortunately, the cost of such elimination is often whose mail seemed suspicious. In many cases, censors prohibitive; in this case, the Light Pink Book recommends even randomly deleted innocent-looking sentences or techniques to limit the bandwidth of covert channels and entire paragraphs in order to prevent secret messages requires auditing to detect any use of the covert channel. A from being delivered. natural question that arises from this suggestion is whether More recently, there has been a great deal of interest in it is feasible for an auditor to do so. digital steganography, that is, in hiding secret messages in This paper focuses on the dual problem of steganogra- communications between computers. This interest is ap- phy: How can two communicating entities send secret parently fueled by the increased amount of communication messages over a public or audited channel so that a third that is mediated by computers and by connections to party such as the reference monitor cannot detect the potential commercial applications: Hidden information presence of the secret messages? Notice how the goal of could potentially be used to detect or limit the unauthorized steganography is different from classical encryption, which propagation of the innocent-looking “carrier” data. Because seeks to conceal the content of secret messages: Stegano- of this, there have been numerous proposals for protocols to graphy is about hiding the very existence of the secret hide data in channels containing pictures [4], [5], video [5], messages. [6], [7], audio [8], [9], and even typeset text [10]. Many of these protocols are extremely clever and rely heavily on . N. Hopper is with the Department of Computer Science and Engineering, domain-specific properties of these channels. On the other University of Minnesota, Minneapolis, MN 55455. hand, the literature on steganography also contains many E-mail: [email protected]. clever attacks that detect the use of such protocols. As a . L. von Ahn is with the Computer Science Department, Carnegie Mellon result, it is unclear from this body of work whether secure University, Pittsburgh, PA 15213. E-mail: [email protected]. J. Langford is with the Machine Learning Group, Yahoo! Research, New steganography is possible at all. York, NY 10018. E-mail: [email protected]. In this paper, we use techniques from cryptography and Manuscript received 26 Apr. 2007; revised 18 July 2008; accepted 15 Sept. complexity theory to answer the question “under what 2008; published online 23 Oct. 2008. conditions is (secure) steganography possible?” We give Recommended for acceptance by F. Lombardi. cryptographic definitions for symmetric-key stegosystems For information on obtaining reprints of this article, please send e-mail to: [email protected], and reference IEEECS Log Number TC-2007-04-0146. and steganographic secrecy against a passive adversary in Digital Object Identifier no. 10.1109/TC.2008.199. terms of indistinguishability from a probabilistic channel 0018-9340/09/$25.00 ß 2009 IEEE Published by the IEEE Computer Society 2 IEEE TRANSACTIONS ON COMPUTERS, VOL. 58, NO. X, XXX 2009 process. We show that a widely believed complexity- bound is weaker than ours (in terms of bits per oracle query) theoretic assumption (the existence of a one-way function) by a factor of log2 2e but applies to a more general class of and access to a channel oracle are both necessary and stegosystems and is proved using techniques of independent sufficient conditions for the existence of secure stegano- interest. Van Le and Kurosawa [23] proposed a stegosystem graphy relative to any channel. We furthermore give a based on arithmetic coding that uses specialized knowledge construction that has essentially optimal bandwidth when of a channel to achieve higher bandwidth per symbol in compared with known provably secure constructions. some cases. Lysyanskaya and Meyerovich [24] considered Finally, we consider the question of robust steganography the effects of an imperfect channel sampling procedure on that resists attempts to censor the use of a covert channel; universal stegosystems. Backes and Cachin [25] introduced a we prove necessary and sufficient conditions for the notion of “active attacks” that use a decoding oracle to detect existence of a secure robust stegosystem and give a the presence of hidden messages in a special “challenge” provably robust stegosystem with nearly optimal band- cover message. width under attack. 1.2 Preliminaries 1.1 Previous Work We denote the length of a string or sequence s by s . We j j The scientific study of steganography in the open literature denote the empty string or sequence by " and the began in 1983 when Simmons [11] stated the problem in terms concatenation of strings s and s by s s . We assume the 1 2 1k 2 of communication in a prison. In his formulation, two use of efficient and unambiguous pairing and unpairing inmates, Alice and Bob, are trying to hatch an escape plan. operators on strings so that s1;s2 may be uniquely The only way they can communicate with each other is ð Þ interpreted as the pairing of s1 with s2 and may not be through a public channel, which is carefully monitored by the the same as s s . One example of such an operation is to 1k 2 warden of the prison, Ward. If Ward detects any encrypted encode s ;s by a prefix-free encoding of s , followed by ð 1 2Þ j 1j messages or codes, he will throw both Alice and Bob into s , followed by a prefix-free encoding of s , and then s . 1 j 2j 2 solitary confinement. The problem of steganography is then: Unpairing works in the obvious way. how can Alice and Bob cook up an escape plan by We let U denote the uniform distribution on 0; 1 k. If X k f g communicating over the public channel in such a way that is a finite set, we denote by x X the action of uniformly Ward does not suspect that anything “unusual” is going on? choosing x from X. We denote by L;l the uniform Anderson and Petitcolas [12], [13] posed many of the L F l distribution on functions f : 0; 1 0; 1 . For a prob- open problems resolved in this article. In particular, they f g !f g ability distribution D, we denote the support of D by D . pointed out that it was unclear how to prove the security of ½ For an integer n, we let n denote the set 1; 2; ...;n . a steganographic protocol and gave an example that is ½ f g similar to the protocol we present in Section 4 but noted that We make use of two different measures of the informa- it was not clear what properties were necessary to prove its tion in a probability distribution. Let be a probability D security. They also posed the open question of bounding the distribution with finite support D; then, the Shannon bandwidth that can be securely achieved over a given cover Entropy of is HS d D Pr d log2 Pr d , and the channel. D ðDÞ ¼ À 2 D½ D½ Minimum Entropy of is H mind D log2 Pr d . D P1ðDÞ ¼ 2 fÀ D½ g Since then, several works [14], [15], [16], [17] have We model all parties by Probabilistic Turing Machines addressed information-theoretic definitions of steganogra- (PTMs).
Recommended publications
  • Reproducibility and Pseudo-Determinism in Log-Space
    Reproducibility and Pseudo-determinism in Log-Space by Ofer Grossman S.B., Massachusetts Institute of Technology (2017) Submitted to the Department of Electrical Engineering and Computer Science in partial fulfillment of the requirements for the degree of Master of Science in Electrical Engineering and Computer Science at the MASSACHUSETTS INSTITUTE OF TECHNOLOGY May 2020 c Massachusetts Institute of Technology 2020. All rights reserved. Author...................................................................... Department of Electrical Engineering and Computer Science May 15, 2020 Certified by.................................................................. Shafi Goldwasser RSA Professor of Electrical Engineering and Computer Science Thesis Supervisor Accepted by................................................................. Leslie A. Kolodziejski Professor of Electrical Engineering and Computer Science Chair, Department Committee on Graduate Students 2 Reproducibility and Pseudo-determinism in Log-Space by Ofer Grossman Submitted to the Department of Electrical Engineering and Computer Science on May 15, 2020, in partial fulfillment of the requirements for the degree of Master of Science in Electrical Engineering and Computer Science Abstract Acuriouspropertyofrandomizedlog-spacesearchalgorithmsisthattheiroutputsareoften longer than their workspace. This leads to the question: how can we reproduce the results of a randomized log space computation without storing the output or randomness verbatim? Running the algorithm again with new
    [Show full text]
  • FOCS 2005 Program SUNDAY October 23, 2005
    FOCS 2005 Program SUNDAY October 23, 2005 Talks in Grand Ballroom, 17th floor Session 1: 8:50am – 10:10am Chair: Eva´ Tardos 8:50 Agnostically Learning Halfspaces Adam Kalai, Adam Klivans, Yishay Mansour and Rocco Servedio 9:10 Noise stability of functions with low influences: invari- ance and optimality The 46th Annual IEEE Symposium on Elchanan Mossel, Ryan O’Donnell and Krzysztof Foundations of Computer Science Oleszkiewicz October 22-25, 2005 Omni William Penn Hotel, 9:30 Every decision tree has an influential variable Pittsburgh, PA Ryan O’Donnell, Michael Saks, Oded Schramm and Rocco Servedio Sponsored by the IEEE Computer Society Technical Committee on Mathematical Foundations of Computing 9:50 Lower Bounds for the Noisy Broadcast Problem In cooperation with ACM SIGACT Navin Goyal, Guy Kindler and Michael Saks Break 10:10am – 10:30am FOCS ’05 gratefully acknowledges financial support from Microsoft Research, Yahoo! Research, and the CMU Aladdin center Session 2: 10:30am – 12:10pm Chair: Satish Rao SATURDAY October 22, 2005 10:30 The Unique Games Conjecture, Integrality Gap for Cut Problems and Embeddability of Negative Type Metrics Tutorials held at CMU University Center into `1 [Best paper award] Reception at Omni William Penn Hotel, Monongahela Room, Subhash Khot and Nisheeth Vishnoi 17th floor 10:50 The Closest Substring problem with small distances Tutorial 1: 1:30pm – 3:30pm Daniel Marx (McConomy Auditorium) Chair: Irit Dinur 11:10 Fitting tree metrics: Hierarchical clustering and Phy- logeny Subhash Khot Nir Ailon and Moses Charikar On the Unique Games Conjecture 11:30 Metric Embeddings with Relaxed Guarantees Break 3:30pm – 4:00pm Ittai Abraham, Yair Bartal, T-H.
    [Show full text]
  • On the Foundations of Cryptography∗
    On the Foundations of Cryptography∗ Oded Goldreich Department of Computer Science Weizmann Institute of Science Rehovot, Israel. [email protected] May 6, 2019 Abstract We survey the main paradigms, approaches and techniques used to conceptualize, define and provide solutions to natural cryptographic problems. We start by presenting some of the central tools used in cryptography; that is, computational difficulty (in the form of one-way functions), pseudorandomness, and zero-knowledge proofs. Based on these tools, we turn to the treatment of basic cryptographic applications such as encryption and signature schemes as well as the design of general secure cryptographic protocols. Our presentation assumes basic knowledge of algorithms, probability theory and complexity theory, but nothing beyond this. Keywords: Cryptography, Theory of Computation. ∗This revision of the primer [59] will appear as Chapter 17 in an ACM book celebrating the work of Goldwasser and Micali. 1 Contents 1 Introduction and Preliminaries 1 1.1 Introduction.................................... ............... 1 1.2 Preliminaries ................................... ............... 4 I Basic Tools 6 2 Computational Difficulty and One-Way Functions 6 2.1 One-WayFunctions................................ ............... 6 2.2 Hard-CorePredicates . .. .. .. .. .. .. .. ................ 9 3 Pseudorandomness 11 3.1 Computational Indistinguishability . ....................... 11 3.2 PseudorandomGenerators. ................. 12 3.3 PseudorandomFunctions . ...............
    [Show full text]
  • Secure Multi-Party Computation in Practice
    SECURE MULTI-PARTY COMPUTATION IN PRACTICE Marcella Christine Hastings A DISSERTATION in Computer and Information Science Presented to the Faculties of the University of Pennsylvania in Partial Fulfillment of the Requirements for the Degree of Doctor of Philosophy 2021 Supervisor of Dissertation Nadia Heninger Adjunct Associate Professor, University of Pennsylvania Associate Professor, University of California, San Diego Graduate Group Chairperson Mayur Naik Professor and Computer and Information Science Graduate Group Chair Dissertation Committee Brett Hemenway Falk, Research Assistant Professor Stephan A. Zdancewic, Professor Sebastian Angel, Raj and Neera Singh Term Assistant Professor abhi shelat, Associate Professor at Khoury College of Computer Sciences, Northeastern University ACKNOWLEDGMENT This dissertation would have been much less pleasant to produce without the presence of many people in my life. I would like to thank my advisor and my dissertation committee for their helpful advice, direction, and long-distance phone calls over the past six years. I would like to thank my fellow PhD students at the University of Pennsylvania, especially the ever-changing but consistently lovely office mates in the Distributed Systems Laboratory and my cohort. Our shared tea-time, cookies, disappointments, and achievements provided an essential community that brought me great joy during my time at Penn. I would like to thank the mentors and colleagues who hosted me at the Security and Privacy Lab at the University of Washington in 2018, the Software & Application Innovation Lab at Boston University in 2019, and the Cryptography and Privacy Research group at Microsoft Research in 2020. My career and happiness greatly benefited from spending these summers exploring fresh research directions and rediscovering the world outside my own work.
    [Show full text]
  • A Decade of Lattice Cryptography
    Full text available at: http://dx.doi.org/10.1561/0400000074 A Decade of Lattice Cryptography Chris Peikert Computer Science and Engineering University of Michigan, United States Boston — Delft Full text available at: http://dx.doi.org/10.1561/0400000074 Foundations and Trends R in Theoretical Computer Science Published, sold and distributed by: now Publishers Inc. PO Box 1024 Hanover, MA 02339 United States Tel. +1-781-985-4510 www.nowpublishers.com [email protected] Outside North America: now Publishers Inc. PO Box 179 2600 AD Delft The Netherlands Tel. +31-6-51115274 The preferred citation for this publication is C. Peikert. A Decade of Lattice Cryptography. Foundations and Trends R in Theoretical Computer Science, vol. 10, no. 4, pp. 283–424, 2014. R This Foundations and Trends issue was typeset in LATEX using a class file designed by Neal Parikh. Printed on acid-free paper. ISBN: 978-1-68083-113-9 c 2016 C. Peikert All rights reserved. No part of this publication may be reproduced, stored in a retrieval system, or transmitted in any form or by any means, mechanical, photocopying, recording or otherwise, without prior written permission of the publishers. Photocopying. In the USA: This journal is registered at the Copyright Clearance Center, Inc., 222 Rosewood Drive, Danvers, MA 01923. Authorization to photocopy items for in- ternal or personal use, or the internal or personal use of specific clients, is granted by now Publishers Inc for users registered with the Copyright Clearance Center (CCC). The ‘services’ for users can be found on the internet at: www.copyright.com For those organizations that have been granted a photocopy license, a separate system of payment has been arranged.
    [Show full text]
  • Hard Communication Channels for Steganography
    Hard Communication Channels for Steganography Sebastian Berndt1 and Maciej Liśkiewicz2 1 University of Lübeck, Lübeck, Germany [email protected] 2 University of Lübeck, Lübeck, Germany [email protected] Abstract This paper considers steganography – the concept of hiding the presence of secret messages in legal communications – in the computational setting and its relation to cryptography. Very re- cently the first (non-polynomial time) steganographic protocol has been shown which, for any communication channel, is provably secure, reliable, and has nearly optimal bandwidth. The security is unconditional, i.e. it does not rely on any unproven complexity-theoretic assumption. This disproves the claim that the existence of one-way functions and access to a communication channel oracle are both necessary and sufficient conditions for the existence of secure steganogra- phy in the sense that secure and reliable steganography exists independently of the existence of one-way functions. In this paper, we prove that this equivalence also does not hold in the more realistic setting, where the stegosystem is polynomial time bounded. We prove this by construct- ing (a) a channel for which secure steganography exists if and only if one-way functions exist and (b) another channel such that secure steganography implies that no one-way functions exist. We therefore show that security-preserving reductions between cryptography and steganography need to be treated very carefully. 1998 ACM Subject Classification E.3 Data Encryption Keywords and phrases provable secure steganography, cryptographic assumptions, pseudoran- dom functions, one-way functions, signature schemes Digital Object Identifier 10.4230/LIPIcs.ISAAC.2016.16 1 Introduction Digital steganography has recently received substantial interest in modern computer science since it allows secret communication without revealing its presence.
    [Show full text]
  • Algebraic Pseudorandom Functions with Improved Efficiency from the Augmented Cascade*
    Algebraic Pseudorandom Functions with Improved Efficiency from the Augmented Cascade* DAN BONEH† HART MONTGOMERY‡ ANANTH RAGHUNATHAN§ Department of Computer Science, Stanford University fdabo,hartm,[email protected] September 8, 2020 Abstract We construct an algebraic pseudorandom function (PRF) that is more efficient than the classic Naor- Reingold algebraic PRF. Our PRF is the result of adapting the cascade construction, which is the basis of HMAC, to the algebraic settings. To do so we define an augmented cascade and prove it secure when the underlying PRF satisfies a property called parallel security. We then use the augmented cascade to build new algebraic PRFs. The algebraic structure of our PRF leads to an efficient large-domain Verifiable Random Function (VRF) and a large-domain simulatable VRF. 1 Introduction Pseudorandom functions (PRFs), first defined by Goldreich, Goldwasser, and Micali [GGM86], are a fun- damental building block in cryptography and have numerous applications. They are used for encryption, message integrity, signatures, key derivation, user authentication, and many other cryptographic mecha- nisms. Beyond cryptography, PRFs are used to defend against denial of service attacks [Ber96, CW03] and even to prove lower bounds in learning theory. In a nutshell, a PRF is indistinguishable from a truly random function. We give precise definitions in the next section. The fastest PRFs are built from block ciphers like AES and security is based on ad-hoc inter- active assumptions. In 1996, Naor and Reingold [NR97] presented an elegant PRF whose security can be deduced from the hardness of the Decision Diffie-Hellman problem (DDH) defined in the next section.
    [Show full text]
  • Omer Reingold, June 2017
    Omer Reingold, June 2017 My research is focused on a wide range of topics within the Foundations of Computer Science. In particular, my core areas, which include my most significant contributions, are Computational Complexity and the Foundations of Cryptography, with an emphasis on Randomness, Derandomization and Explicit Combinatorial Constructions. However, I am interested in a much broader range of topics (some closer and some further away from my core areas). Examples include Differential Privacy and Fairness, Game Theory, Hashing and Data Structures, Resource Allocation and Data Analysis. Current Academic Employment: Professor of Computer Science, Stanford University, Since Feb. 2015. Education: • 1994-1998 Weizmann Institute of Science, Rehovot, Israel. Ph.D. in Computer Science. Thesis title: Pseudo-random synthesizers, functions and permutations. Advisor: Prof. Moni Naor. • 1991-1994 Tel-Aviv University, Tel-Aviv, Israel. B.Sc in Mathematics and Computer Science - Summa Cum Laude. Selected Awards and distinctions (additional awards below): • The 2009 Gödel Prize for outstanding papers in the area of theoretical computer science (jointly with Salil Vadhan and Avi Wigderson). • 2005 ACM Grace Murray Hopper Award. Awarded by the Association for Computing Machinery (ACM) to “the outstanding young computer professional of the year.” • ACM Fellow (2014) Previous Academic Employment and Short Term Visits: • Feb 2015 – September 2016 Principal Research Engineer, Samsung Research America (SRA) • Nov 2014 – Feb 2015 (while on leave from
    [Show full text]
  • Draft Lecture Notes
    CS-282A / MATH-209A Foundations of Cryptography Draft Lecture Notes Winter 2010 Rafail Ostrovsky UCLA Copyright © Rafail Ostrovsky 2003-2010 Acknowledgements: These lecture notes were produced while I was teaching this graduate course at UCLA. After each lecture, I asked students to read old notes and update them according to the lecture. I am grateful to all of the students who participated in this massive endeavor throughout the years, for asking questions, taking notes and improving these lecture notes each time. Table of contents PART 1: Overview, Complexity classes, Weak and Strong One-way functions, Number Theory Background. PART 2: Hard-Core Bits. PART 3: Private Key Encryption, Perfectly Secure Encryption and its limitations, Semantic Security, Pseudo-Random Generators. PART 4: Implications of Pseudo-Random Generators, Pseudo-random Functions and its applications. PART 5: Digital Signatures. PART 6: Trapdoor Permutations, Public-Key Encryption and its definitions, Goldwasser-Micali, El-Gamal and Cramer-Shoup cryptosystems. PART 7: Interactive Proofs and Zero-Knowledge. PART 8: Bit Commitment Protocols in different settings. PART 9: Non-Interactive Zero Knowledge (NIZK) PART 10: CCA1 and CCA2 Public Key Encryption from general Assumptions: Naor-Yung, DDN. PART 11: Non-Malleable Commitments, Non-Malleable NIZK. PART 12: Multi-Server PIR PART 13: Single-Server PIR, OT, 1-2-OT, MPC, Program Obfuscation. PART 14: More on MPC including GMW and BGW protocols in the honest-but-curious setting. PART 15: Yao’s garbled circuits, Efficient ZK Arguments, Non-Black-Box Zero Knowledge. PART 16: Remote Secure Data Storage (in a Cloud), Oblivious RAMs. CS 282A/MATH 209A: Foundations of Cryptography °c 2006-2010 Prof.
    [Show full text]
  • How to Encipher Messages on a Small Domain Deterministic Encryption and the Thorp Shuffle
    How to Encipher Messages on a Small Domain Deterministic Encryption and the Thorp Shuffle Ben Morris1, Phillip Rogaway2, and Till Stegers2 1 Dept. of Mathematics, University of California, Davis, California 95616, USA 2 Dept. of Computer Science, University of California, Davis, California 95616, USA Abstract. We analyze the security of the Thorp shuffle, or, equivalently, a maximally unbalanced Feistel network. Roughly said, the Thorp shuffle on N cards mixes any N 1−1/r of them in O(r lg N) steps. Correspond- ingly, making O(r) passes of maximally unbalanced Feistel over an n-bit string ensures CCA-security to 2n(1−1/r) queries. Our results, which em- ploy Markov-chain techniques, enable the construction of a practical and provably-secure blockcipher-based scheme for deterministically encipher- ing credit card numbers and the like using a conventional blockcipher. 1 Introduction Small-space encryption. Suppose you want to encrypt a 9-decimal-digit plaintext, say a U.S. social-security number, into a ciphertext that is again a 9-decimal-digit number. A shared key K is used to control the encryption. Syn- tactically, you seek a cipher E: K×M → M where M = {0, 1,...,N− 1}, 9 N =10 ,andEK = E(K, ·) is a permutation for each key K ∈K. You aim to construct your scheme from a well-known primitive, say AES, and to prove your scheme is as secure as the primitive from which you start. The problem is harder than it sounds. You can’t just encode each plaintext M ∈Mas a 128-bit string and then apply AES, say, as that will return a 128-bit string and projecting back onto M will destroy permutivity.
    [Show full text]
  • Immunizing Encryption Schemes from Decryption Errors
    Immunizing Encryption Schemes from Decryption Errors Cynthia Dwork1, Moni Naor2, and Omer Reingold2 1 Microsoft Research, SVC 1065 L’Avenida Mountain View, CA 94043 [email protected] 2 Weizmann Institute of Science Rehovot 76100, Israel {moni.naor,omer.reingold}@weizmann.ac.il Abstract. We provide methods for transforming an encryption scheme susceptible to decryption errors into one that is immune to these errors. Immunity to decryption errors is vital when constructing non-malleable and chosen ciphertext secure encryption schemes via current techniques; in addition, it may help defend against certain cryptanalytic techniques, such as the attack of Proos [33] on the NTRU scheme. When decryption errors are very infrequent, our transformation is ex- tremely simple and efficient, almost free. To deal with significant error probabilities, we apply amplification techniques translated from a re- lated information theoretic setting. These techniques allow us to correct even very weak encryption schemes where in addition to decryption er- rors, an adversary has substantial probability of breaking the scheme by decrypting random messages (without knowledge of the secret key). In other words, under these weak encryption schemes, the only guaranteed difference between the legitimate recipient and the adversary is in the frequency of decryption errors. All the above transformations work in a standard cryptographic model; specifically, they do not rely on a random oracle. We also consider the random oracle model, where we give a simple trans- formation from a one-way encryption scheme which is error-prone into one that is immune to errors. We conclude that error-prone cryptosystems can be used in order to create more secure cryptosystems.
    [Show full text]
  • Computational Hardness of Optimal Fair Computation: Beyond Minicrypt
    Computational Hardness of Optimal Fair Computation: Beyond Minicrypt Hemanta K. Maji Department of Computer Science, Purdue University, USA [email protected] Mingyuan Wang Department of Computer Science, Purdue University, USA [email protected] Abstract Secure multi-party computation allows mutually distrusting parties to compute securely over their private data. However, guaranteeing output delivery to honest parties when the adversarial parties may abort the protocol has been a challenging objective. As a representative task, this work considers two-party coin-tossing protocols with guaranteed output delivery, a.k.a., fair coin- tossing. In the information-theoretic plain model, as in two-party zero-sum games, one of the parties can force an output with certainty. In the commitment-hybrid, any r-message coin-tossing proto- √ √ col is 1/ r-unfair, i.e., the adversary can change the honest party’s output distribution by 1/ r in the statistical distance. Moran, Naor, and Segev (TCC–2009) constructed the first 1/r-unfair protocol in the oblivious transfer-hybrid. No further security improvement is possible because Cleve (STOC–1986) proved that 1/r-unfairness is unavoidable. Therefore, Moran, Naor, and Segev’s coin-tossing protocol is optimal. However, is oblivious transfer necessary for optimal fair coin-tossing? Maji and Wang (CRYPTO–2020) proved that any coin-tossing protocol using one-way func- √ tions in a black-box manner is at least 1/ r-unfair. That is, optimal fair coin-tossing is impossible in Minicrypt. Our work focuses on tightly characterizing the hardness of computation assump- tion necessary and sufficient for optimal fair coin-tossing within Cryptomania, outside Minicrypt.
    [Show full text]