PL2: Towards Predictable Low Latency in Rack-Scale Networks

Total Page:16

File Type:pdf, Size:1020Kb

PL2: Towards Predictable Low Latency in Rack-Scale Networks PL2: Towards Predictable Low Latency in Rack-Scale Networks Yanfang Le y, Radhika Niranjan Mysore yy, Lalith Suresh yy, Gerd Zellwegeryy, Sujata Banerjeeyy, Aditya Akellay, Michael Swifty University of Wisconsin-Madisony, VMware Researchyy Abstract Rack-scale networks1 need to satisfy the key requirements High performance rack-scale offerings package disaggre- of uniform low latency and high utilization, irrespec- gated pools of compute, memory and storage hardware in a tive of where applications reside, and which accelerators single rack to run diverse workloads with varying require- they access (e.g., FPGA vs. CPU vs. GPU). However, three ments, including applications that need low and predictable key obstacles stand in the way of achieving these goals latency. The intra-rack network is typically high speed Ether- because of the above-mentioned characteristics. First, the net, which can suffer from congestion leading to packet drops rack-scale network must be transport-agnostic, a neces- and may not satisfy the stringent tail latency requirements sity in environments with (a) heterogeneous accelerator de- 2 for some workloads (including remote memory/storage ac- vices that have different characteristics than CPU network cesses). In this paper, we design a Predictable Low Latency stacks [25, 38, 39], and (b) increasing use of CPU-bypass net- (PL2) network architecture for rack-scale systems with Eth- working [32, 35, 37]. Second, Ethernet is not a lossless fabric, ernet as interconnecting fabric. PL2 leverages programmable and yet, our experiments (§2) on a 100G testbed confirm Ethernet switches to carefully schedule packets such that that drops, not queueing, are the largest contributor to tail they incur no loss with NIC and switch queues maintained at latency pathologies. Third, the design must be workload- small, near-zero levels. In our 100 Gbps rack-prototype, PL2 oblivious – given that we cannot anticipate traffic and work- keeps 99th-percentile memcached RPC latencies under 60 µs load patterns across a broad range of customer environments, even when the RPCs compete with extreme offered-loads of it is impractical to rely on state-of-the-art proposals (§3) that 400%, without losing traffic. Network transfers for a machine hinge on configuring rate limits or priorities using a-priori learning training task complete 30% faster than a receiver- knowledge of the workload. driven scheme implementation modelled after Homa (222ms In this paper, we present Predictable Low Latency or PL2, vs 321ms 99%ile latency per iteration). a rack-scale lossless network architecture that achieves low latency and high throughput in a transport-agnostic and workload-oblivious manner. PL2 reduces NIC-to-NIC laten- cies by proactively avoiding losses. PL2 supports a variety of 1 Introduction message transport protocols and gracefully accommodates increasing numbers of flows, even at 100G line rates. It nei- Rack-scale data center solutions like Dell-EMC VxRail [57] ther requires a-priori knowledge of workload characteristics and Intel RSD [52] have emerged as a new building block nor depends on rate-limits or traffic priorities to be set based for modern enterprise, cloud, and edge infrastructure. These on workload characteristics. rack-units have three key characteristics; First is the increas- arXiv:2101.06537v2 [cs.NI] 22 Jan 2021 To achieve these goals, senders in PL2 explicitly request a ing use of resource disaggregation and hardware accelerators switch buffer reservation for a given number of packets, a within these rack-units like GPUs and FPGAs [52], in ad- packet burst, and receive notification as to when that burst dition to high-density compute and storage units. Second, can be transmitted without facing any cross traffic from other Ethernet is by far the dominant interconnect of choice within senders. PL2 achieves this form of centralized scheduling such racks, even for communication between compute units, even at 100G line rates by overcoming the key challenge storage units and accelerators (e.g., Ethernet-pooled FPGA of carefully partitioning the scheduling responsibility be- and NVMe in Intel RDS [33]). Third, these racks are deployed tween hosts in the rack and the Top-of-Rack (ToR) switch. in a wide range of enterprise and cloud customer environ- In particular, the end-host protocol is kept simple enough ments, running a heterogeneous mix of modern (e.g., ma- chine learning, graph processing) and legacy applications 1The network extending between NICs of such rack-units across the top-of- (e.g., monolithic web applications), making it impractical to rack (ToR) switch. 2For example, FPGA stacks will not be connection-oriented due to scaling anticipate traffic and workload patterns. issues [54] and GPUs will not have a single receiver stack [39]). 1 to accommodate accelerator devices and implementations adds latencies to network transfers. However, the implication within NICs (§4.4), whereas the timeslot allocation itself is of this trend is that microbursts can over-run shared output performed in the ToR switch at line rate (as opposed to doing switch-buffers and cause drops. For instance, a 2 MB buffer so on a host, which is prone to software overheads). in a ToR switch with 100 Gbps ports provides buffering for In summary, our contributions are: just 160 µs which means only 8 simultaneous transfers of • The PL2 design that embodies novel yet simple algo- 2Mbits can be sustained before the switch starts dropping rithms for lossless transmissions and near-zero queu- packets. Today’s rack-scale networks support up to 64 rack- ing within a rack units [4], where each end-system can have tens of thousands • A PL2 implementation using a P4 programmable switch of ongoing transfers. At that scale a 2 MB can be overrun and an end-host stack that leverages Mellanox’s state- with only 6 simultaneous 5 µs (1 RTT) network transfers per of-the-art VMA message acceleration library [7] rack-scale unit. In short, as rack-densities go up, drops due to • A comprehensive PL2 evaluation on a 100 Gbps pro- microbursts can be frequent. Therefore, assumptions made totype, supporting three different transports (TCP, by congestion protocols like [30, 44] that the network-core UDP and Raw Ethernet), all benefiting from near-zero (ToR switch in the case of racks) is lossless, no longer holds. queueing in the network. Compared to VMA, we demon- strate up to 2.2x improvement in the 99th percentile 3. Rack-scale traffic is ON/OFF traffic [19] We believe latency for the Memcached application; a 20% improve- this trend will continue with network traffic generated by ac- ment to run a VGG16 machine learning workload; and celerators. Measuring traffic-demands in such environments near-optimal latency and throughput in experiments is hard, let alone learning about workload-characteristics; using trace-based workload generators. traffic demands at ns-timescales will be different compared to `s timescales and ms-timescales [59]. Workload churn and 2 Motivation different application mixes adds to the unpredictability. The primary goal of PL2 is to provide uniformly low-latency Coming up with effective rate-limits18 [ , 27, 34, 36, 41], across Ethernet rack-fabrics, while achieving high-utilization. and readjusting these rate-limits with changing traffic-conditions We take inspiration from prior work around low and pre- in time (i.e., less than an RTT) is impossible; so is setting dictable latency within data center networks [13, 15, 16, 18, relative packet priorities [44] effectively [40] so that impor- 20, 22, 26, 27, 29–31, 34, 36, 41–44, 47, 48, 56, 58, 60], but find tant traffic is never lost or delayed. In our experience nei- that rack-scale networks provide a rich set of new challenges. ther rate-limits nor priority prescription is an answer to Rack-scale characteristics and implications congestion-control within a rack. 1. Even as line-rates increase, intra-rack RTTs are Drops cause the most noticeable tails not getting smaller. [44] measured 5 µs end-to-end RTT Based on the above three observations, we believe that new on a 10 Gbps testbed with a single ToR switch, inclusive techniques are required to ensure low-latency and high- of software delays on the servers. A 64B packet still has utilization within a rack-scale network. We hinge the PL2 an RTT of 5 µs in our 100 Gbps rack-scale PL2 prototype. design on the observation that drops, not queuing cause the Even though the network transmission times reduce propor- most noticeable tails. tionally with increasing line-rates, switching-delays, NIC We illustrate this point with an experiment that introduces hardware delays, and DMA transfer delays at end-hosts have microbursts in our 100 Gbps testbed even when the network remained about the same, and these delays together domi- is partially loaded, by using 5-way incast of roughly 18 Gbps nate transmission delays in a rack. Forward-error-correction traffic per sender. All messages transmitted in our exper- delays increase as line-rates increase, and can add variability iment can be transmitted within a single RTT. As shown of up to 1-2 µs in a 100 Gbps network. This implies that rack- in Figure 1a, the 99%ile latencies experienced by a receiver- scale networks are able to transfer more data in the same RTT driven scheme (RDS-baseline) based on Homa (described as interconnect speeds increase. Flows up to 61 kB can com- in detail in Section 6) correspond to the maximum output- plete within 1 RTT with a 100 Gbps backplane as opposed to buffering available in the ToR switch (around 120 µs in Fig- 6 kB for a 10 Gbps backplane. For congestion control proto- ure 1b), while the 99.9%ile latencies correspond to delays due cols and predictable latency schemes to be effective for flows to drops, and are two orders of magnitude higher.
Recommended publications
  • IPFS and Friends: a Qualitative Comparison of Next Generation Peer-To-Peer Data Networks Erik Daniel and Florian Tschorsch
    1 IPFS and Friends: A Qualitative Comparison of Next Generation Peer-to-Peer Data Networks Erik Daniel and Florian Tschorsch Abstract—Decentralized, distributed storage offers a way to types of files [1]. Napster and Gnutella marked the beginning reduce the impact of data silos as often fostered by centralized and were followed by many other P2P networks focusing on cloud storage. While the intentions of this trend are not new, the specialized application areas or novel network structures. For topic gained traction due to technological advancements, most notably blockchain networks. As a consequence, we observe that example, Freenet [2] realizes anonymous storage and retrieval. a new generation of peer-to-peer data networks emerges. In this Chord [3], CAN [4], and Pastry [5] provide protocols to survey paper, we therefore provide a technical overview of the maintain a structured overlay network topology. In particular, next generation data networks. We use select data networks to BitTorrent [6] received a lot of attention from both users and introduce general concepts and to emphasize new developments. the research community. BitTorrent introduced an incentive Specifically, we provide a deeper outline of the Interplanetary File System and a general overview of Swarm, the Hypercore Pro- mechanism to achieve Pareto efficiency, trying to improve tocol, SAFE, Storj, and Arweave. We identify common building network utilization achieving a higher level of robustness. We blocks and provide a qualitative comparison. From the overview, consider networks such as Napster, Gnutella, Freenet, BitTor- we derive future challenges and research goals concerning data rent, and many more as first generation P2P data networks, networks.
    [Show full text]
  • A Fog Storage Software Architecture for the Internet of Things Bastien Confais, Adrien Lebre, Benoît Parrein
    A Fog storage software architecture for the Internet of Things Bastien Confais, Adrien Lebre, Benoît Parrein To cite this version: Bastien Confais, Adrien Lebre, Benoît Parrein. A Fog storage software architecture for the Internet of Things. Advances in Edge Computing: Massive Parallel Processing and Applications, IOS Press, pp.61-105, 2020, Advances in Parallel Computing, 978-1-64368-062-0. 10.3233/APC200004. hal- 02496105 HAL Id: hal-02496105 https://hal.archives-ouvertes.fr/hal-02496105 Submitted on 2 Mar 2020 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. November 2019 A Fog storage software architecture for the Internet of Things Bastien CONFAIS a Adrien LEBRE b and Benoˆıt PARREIN c;1 a CNRS, LS2N, Polytech Nantes, rue Christian Pauc, Nantes, France b Institut Mines Telecom Atlantique, LS2N/Inria, 4 Rue Alfred Kastler, Nantes, France c Universite´ de Nantes, LS2N, Polytech Nantes, Nantes, France Abstract. The last prevision of the european Think Tank IDATE Digiworld esti- mates to 35 billion of connected devices in 2030 over the world just for the con- sumer market. This deep wave will be accompanied by a deluge of data, applica- tions and services.
    [Show full text]
  • Decentralized File Sharing Application for Mobile Devices -. | Davide Taibi
    Asterism: Decentralized File Sharing Application for Mobile Devices Olli-Pekka Heinisuo Valentina Lenarduzzi Davide Taibi Tampere University Tampere University Tampere University Tampere, Finland Tampere, Finland Tampere, Finland [email protected] valentina.lenarduzzi@tut.fi davide.taibi@tut.fi Abstract—Most applications and services rely on central au- To avoid the issues of centralization, data can to be stored thorities. This introduces a single point of failure to the system. without any central authorities. Decentralized storage and The central authority must be trusted to have data stored by the content distribution can be achieved with peer-to-peer (P2P) application available at any given time. More importantly, the privacy of the user depends on the service provider capability to networking. In peer-to-peer networking every node of a network keep the data safe. A decentralized system could be a solution to is both a client and a server [4]. The nodes then talk to each remove the dependency from a central authority. Moreover, due other without any central service as opposed to centralized to the rapid growth of mobile device usage, the availability of systems where all communication goes through dedicated decentralization must not be limited only to desktop computers. central servers. In this work we aim at studying the possibility to use mobile devices as a decentralized file sharing platform without any Peer-to-peer (P2P networks and file sharing are rather central authorities. This was done by implementing Asterism, common on desktop environments but they have been rarely a peer-to-peer file-sharing mobile application based on the Inter- used on mobile devices due to more restricted resources.
    [Show full text]
  • Forensic Investigation of P2P Cloud Storage Services and Backbone For
    Forensic investigation of P2P cloud storage services and backbone for IoT networks : BitTorrent Sync as a case study Yee Yang, T, Dehghantanha, A, Choo, R and Yang, LK http://dx.doi.org/10.1016/j.compeleceng.2016.08.020 Title Forensic investigation of P2P cloud storage services and backbone for IoT networks : BitTorrent Sync as a case study Authors Yee Yang, T, Dehghantanha, A, Choo, R and Yang, LK Type Article URL This version is available at: http://usir.salford.ac.uk/id/eprint/40497/ Published Date 2016 USIR is a digital collection of the research output of the University of Salford. Where copyright permits, full text material held in the repository is made freely available online and can be read, downloaded and copied for non-commercial private study or research purposes. Please check the manuscript for any further copyright restrictions. For more information, including our policy and submission procedure, please contact the Repository Team at: [email protected]. Note: This is authors accepted copy – for final article please refer to International Journal of Computers & Electrical Engineering Forensic Investigation of P2P Cloud Storage: BitTorrent Sync as a Case Study 1 2 3 1 Teing Yee Yang , Ali Dehghantanha , Kim-Kwang Raymond Choo , Zaiton Muda 1 Department of Computer Science, Faculty of Computer Science and Information Technology, Universiti Putra Malaysia, UPM Serdang, Selangor, Malaysia 2 The School of Computing, Science & Engineering, Newton Building, University of Salford, Salford, Greater Manchester, United Kingdom 3 Information Assurance Research Group, University of South Australia, Adelaide, South Australia, Australia. Abstract Cloud computing has been regarded as the technology enabler for the Internet of Things (IoT).
    [Show full text]
  • Download-Emule-Kad-Server-List.Pdf
    Download Emule Kad Server List Download Emule Kad Server List 1 / 3 2 / 3 web site page displaying list of all active servers on the eDonkey/eMule p2p network. ... ping test update servers list at client start download list in eMule.. 0.50a installed on my computer. I can connect to eD2K network easily but I can't connect to Kad network. I have tried to download from http://www.nodes-dat.com/ but the first button " Add to eMule (from Nodes Server)" did't work and the other two worked but the problem still remains.. Bezpieczna lista serwerów emule do pobrania. Pobierz listę zawsze aktualną. Download server.met & serverlist for eMule.. eMule now connects to both the eDonkey network and the Kad network. ... eMule will use clients it knows already from the ed2k servers to get connected to Kad .... The servers merely help hold the network together. Meanwhile, Kad is a network that is also connectable via eMule. Unlike the ED2K network, ... You can use the easy to use installer or you can download the binaries. The difference is that the .... nodes.dat nodes for emule kademlia net server edonkey overnet. ... von IP/Port im Kad-Fenster, oder. - per Download aus dem Internent, z.B. nodes.dat.. Connecting to servers hasn't been working for a long time. ... started them again (there is free drive space on the download drive) but can't get a Kad connection. ... Block4: ipfilter.dat, nodes.dat, server.met (emule-security.org). Dodaj do #eMule te 2 pliki : Do serwerów, czyli eD2k --- http://www.server-list.info/ Do Kad ---..
    [Show full text]
  • Content Addressed, Versioned, P2P File System (DRAFT 3)
    IPFS - Content Addressed, Versioned, P2P File System (DRAFT 3) Juan Benet [email protected] ABSTRACT parties invested in the current model. But from another per- The InterPlanetary File System (IPFS) is a peer-to-peer dis- spective, new protocols have emerged and gained wide use tributed file system that seeks to connect all computing de- since the emergence of HTTP. What is lacking is upgrading vices with the same system of files. In some ways, IPFS design: enhancing the current HTTP web, and introducing is similar to the Web, but IPFS could be seen as a sin- new functionality without degrading user experience. gle BitTorrent swarm, exchanging objects within one Git Industry has gotten away with using HTTP this long be- repository. In other words, IPFS provides a high through- cause moving small files around is relatively cheap, even for put content-addressed block storage model, with content- small organizations with lots of traffic. But we are enter- addressed hyper links. This forms a generalized Merkle ing a new era of data distribution with new challenges: (a) DAG, a data structure upon which one can build versioned hosting and distributing petabyte datasets, (b) computing file systems, blockchains, and even a Permanent Web. IPFS on large data across organizations, (c) high-volume high- combines a distributed hashtable, an incentivized block ex- definition on-demand or real-time media streams, (d) ver- change, and a self-certifying namespace. IPFS has no single sioning and linking of massive datasets, (e) preventing ac- point of failure, and nodes do not need to trust each other.
    [Show full text]
  • African Development Bank Group Mali
    AFRICAN DEVELOPMENT BANK GROUP Public Disclosure Authorized Disclosure Public MALI PROJECT FOR THE ECONOMIC EMPOWERMENT OF WOMEN IN THE SHEA BUTTER SUBSECTOR (PAEFFK) Public Disclosure Authorized Disclosure Public RDGW DEPARTMENT November 2018 Translated document TABLE OF CONTENTS Currency Equivalents and Fiscal Year ................................................................................................................i Acronyms and Abbreviations ............................................................................................................................ii Project Brief ................................................................................................................................................... iii Project summary............................................................................................................................................... v Results-Based Logical Framework ................................................................................................................... vi Project Implementation Schedule ……………………………………………………………………………......viii I. STRATEGIC THRUST AND RATIONALE ............................................................................................ 1 1.1. Project Linkages with the Country Strategy and Objectives .................................................................. 1 1.2. Rationale for Bank Involvement ........................................................................................................... 1 1.3. Aid Coordination
    [Show full text]
  • A Distributed Content Addressable Network for Open Educational Resources
    DOI: 10.33965/celda2019_201911C063 16th International Conference on Cognition and Exploratory Learning in Digital Age (CELDA 2019) A DISTRIBUTED CONTENT ADDRESSABLE NETWORK FOR OPEN EDUCATIONAL RESOURCES Stephen Downes National Research Council Canada, Ottawa, Ontario, Canada ABSTRACT We introduce Content Addressable Resources for Education (CARE) as a method for addressing issues of scale, access, management and distribution that currently exist for open educational resources (OER) as they are currently developed in higher education. CARE is based on the concept of the distributed web (dweb) and, using (for example) the Interplanetary File System (IPFS) provides a means to distributed OER in such a way that they cannot be blocked or paywalled, can be associated with each other (for example, as links in a single site, or as newer versions of existing resources) creating what is essentially an open resource graph (ORG), and when accessed through applications such as Beaker Browser, can be cloned and edited by any user to create and share new resources. KEYWORDS Open Educational Resources, Distributed Web, Dat, Interplanetary File System, Open Resource Graph, Content-based Addressing 1. INTRODUCTION The distributed web potentially solves issues proponents of decentralization have long sought to address. One challenge is traffic, which overloads a single server. A second issue is latency, or the lag created by accessing resources half a world away. Additionally, some resources may be subject to national policies creating the need to differentiate access. And finally, if the centralized source is unavailable for some reason, then access for the entire world is disrupted. For the World Wide Web, many of these challenges are addressed with Content Distribution Networks (CDN).
    [Show full text]
  • An Efficient Holistic Data Distribution and Storage Solution for Online Social Networks Guoxin Liu Clemson University
    Clemson University TigerPrints All Dissertations Dissertations 12-2015 An Efficient Holistic Data Distribution and Storage Solution for Online Social Networks Guoxin Liu Clemson University Follow this and additional works at: https://tigerprints.clemson.edu/all_dissertations Recommended Citation Liu, Guoxin, "An Efficient Holistic Data Distribution and Storage Solution for Online Social Networks" (2015). All Dissertations. 1774. https://tigerprints.clemson.edu/all_dissertations/1774 This Dissertation is brought to you for free and open access by the Dissertations at TigerPrints. It has been accepted for inclusion in All Dissertations by an authorized administrator of TigerPrints. For more information, please contact [email protected]. An Efficient Holistic Data Distribution and Storage Solution for Online Social Networks A Dissertation Presented to the Graduate School of Clemson University In Partial Fulfillment of the Requirements for the Degree Doctor of Philosophy Computer Engineering by Guoxin Liu December 2015 Accepted by: Dr. Haiying Shen, Committee Chair Dr. Richard R. Brooks Dr. Kuang-Ching (KC) Wang Dr. Feng Luo Abstract In the past few years, Online Social Networks (OSNs) have dramatically spread over the world. Facebook [4], one of the largest worldwide OSNs, has 1.35 billion users, 82.2% of whom are outside the US [36]. The browsing and posting interactions (text content) between OSN users lead to user data reads (visits) and writes (updates) in OSN datacenters, and Facebook now serves a billion reads and tens of millions of writes per second [37]. Besides that, Facebook has become one of the top Internet traffic sources [36] by sharing tremendous number of large multimedia files including photos and videos.
    [Show full text]
  • Bittorrent App Change Download Location How to Seed Moved Or Renamed Files in Bittorrent
    bittorrent app change download location How to Seed Moved or Renamed Files in BitTorrent. BitTorrent etiquette dictates that once you've downloaded files, you also seed them to help others' downloads go quickly. If you need to move or rename your download, everything's thrown off. Reader Jake Champion explains how to seed after moving or renaming files. This feature is available in uTorrent version 1.8 and later. You can not change file names and keep seeding with earlier versions of uTorrent. You can use this guide if you move or rename the folder in addition to changing file names. It is possible to change file (and folder) names because they are not really parts of the files. They are like ''labels'' that tell the computer where to find the content of the files. After such changes, you need to retarget the files in uTorrent if you want to keep seeding the torrent. The procedure for doing this is the same no matter if the torrent is already loaded in uTorrent, or if you add it as a new torrent. Seed Renamed Files. 1. If the torrent is already loaded in uTorrent, right-click it and choose ''Stop'' (if it is not already stopped). If you are adding a new torrent with changed file names, make sure that you add it in stopped mode (or stop it very quickly if you fail to do this). 2. Select the torrent, then click the ''Files'' tab in the lower pane. (If you have no lower pane, click ''Options'' and check ''Show detailed info''.) 3.
    [Show full text]
  • Solid Over the Interplanetary File System
    Solid over the Interplanetary File System Fabrizio Parrillo Christian Tschudin Dept of Mathematics and Computer Science Dept of Mathematics and Computer Science University of Basel, Switzerland University of Basel, Switzerland <[email protected]> <[email protected]> Identity Providers Service Identity Providers Service Abstract—Solid is a moderate re-decentralization architecture Clients for the Web. In Solid, applications fetch data from storage Provider Provider providers called “pods” which implement discretionary access control, permitting users to grant and revoke access rights and HTTP over the Internet HTTP over the Internet thus keeping control over their data. In this paper, we propose Pod Providers Pod Providers to apply an additional de-verticalization by introducing IPFS as File, SPARQL, File, SPARQL, File, SPARQL, File, SPARQL, a common backend for Solid pods, the goal being to prevent pod Access Ctrl Access Ctrl Access Ctrl Access Ctrl provider lock-in and permitting users to consider “self-poding” Perssistent Perssistent IPFS backend where the heavy persistency task is out-sourced to IPFS. We have implemented this Solid-over-IPFS architecture using the open- source “Community Solid Server” and successfully ran several Fig. 1. Replacing Solid’s persistency layer with IPFS Solid applications over IPFS. Index Terms—IPFS, Solid, Dweb (see Fig. 1). In this way, IPFS becomes the uniting element I. INTRODUCTION among all Solid pod operators, effectively preventing them The Solid project [1] is Tim Berners Lee’s answer to the from holding the data hostage. Moreover, end-users and their massive centralization that has occurred in the Web: a few browser do not have to (but can) be IPFS-aware because this companies dominate the Web ecosystem and are safeguarding burden now shifts to the tech-savvy pod operators, offering a and mediating most of the world’s social media content.
    [Show full text]
  • Intro to Crypto Sci-Club Notes
    Intro to Crypto Sci-Club Notes April 25, 2019 Abstract Notes preparing for sci-club crypto-blockchain presentation. List of possible topics Start with history, review of core tech concepts. Finish with applications. • First large-scale use of blockchain was in WWII Enigma coding machines. Maybe explain how these work? • Concept of mixing/entropy. Concept of mixing at different time-scales (slow, fast). • Public key crypto mathematics theory. • Public key crypto examples: PGP/GPG, x.509 certs, SSL/https, onion routing described at general level • Crypto at a detailed level: ratchet, double ratchet, forward secrecy, what is it for, how does this work? • Nonce; using nonces for secret boxes; all secret boxes must have a different nonce else there is no forward secrecy • Merkle tree (hash tree): it doesn’t have to be an actual chain; it can be a tree of chains, leading up to a signed root. Thus the dat:// protocol and btrfs, zfs – Bitcoin has merkle trees inside of it; where? why? • Git as a blockchain, and why it was needed - to prevent source code fraud (real, not imagined - intentional covert, secret insertion of corruption/bugs into Linux kernel by unknown actors (possibly NSA??)) – Git is a blockchain of commits. – Each commit is a Merkle tree of file-objects (blobs) that captures the state of that filesystem tree at that particular moment. • File sharing, sleepycat/napster, 1 – Gnutella, BearShare, LimeWire, Shareaza, all use tiger tree hash • torrent architectures (bittorrent tracker is example of DHT) • Other weirdo crypto ideas: death-lottery. • Forking and fork resolution. A fork prevents maintanance of state (e.g.
    [Show full text]