BASICS of CODING THEORY CODING

Total Page:16

File Type:pdf, Size:1020Kb

BASICS of CODING THEORY CODING CHAPTER 1: BASICS of CODING THEORY ABSTRACT Coding theory - theory of error correcting codes - is one of the most interesting and applied part of informatics. Part I Goals of coding theory are to develop systems and methods that allow to detect/correct errors caused when information is transmitted through noisy channels. Basics of coding theory All real communication systems that work with digitally represented data, as CD players, TV, fax machines, internet, satellites, mobiles, require to use error correcting codes because all real channels are, to some extent, noisy – due to various interference/destruction caused by the environment Coding theory problems are therefore among the very basic and most frequent problems of storage and transmission of information. Coding theory results allow to create reliable systems out of unreliable systems to store and/or to transmit information. Coding theory methods are often elegant applications of very basic concepts and methods of (abstract) algebra. This first chapter presents and illustrates the very basic problems, concepts, methods and results of coding theory. prof. Jozef Gruska IV054 1. Basics of coding theory 2/51 CODING - BASIC CONCEPTS CHANNEL Without coding theory and error-correcting codes there would be no deep-space travel is any physical medium in which information is stored or through which information is and pictures, no satellite TV, no compact disc, no . no . no . transmitted. Error-correcting codes are used to correct messages when they are (erroneously) (Telephone lines, optical fibres and also the atmosphere are examples of channels.) transmitted through noisy channels. NOISE channel message code code may be caused by sunspots, lighting, meteor showers, random radio disturbance, poor W Encoding word word Decoding W user source typing, poor hearing, . C(W) C'(W)noise TRANSMISSION GOALS Error correcting framework 1 Fast encoding of information. Example 2 Similar messages should be encoded very differently 3 Easy transmission of encoded messages. message Encoding 00000 01001 Decoding YES YES 00000 01001 YES user YES or NO 4 Fast decoding of received messages. NO 11111 00000 5 Reliable correction of errors introduced in the channel. A code C over an alphabet Σ is a subset of Σ∗(C Σ∗). ⊆ 6 Maximum transfer of information per unit time. A q-nary code is a code over an alphabet of q-symbols. A binary code is a code over the alphabet 0, 1 . BASIC METHOD OF FIGHTING ERRORS: REDUNDANCY!!! { } Examples of codes C1 = 00, 01, 10, 11 C2 = 000, 010, 101, 100 { } { } C3 = 00000, 01101, 10111, 11011 Example: 0 is encoded as 00000 and 1 is encoded as 11111. { } prof. Jozef Gruska IV054 1. Basics of coding theory 3/51 prof. Jozef Gruska IV054 1. Basics of coding theory 4/51 CHANNELS - MATHEMATICAL VIEWS BASIC CHANNEL CODING PROBLEMS Summary: The task of a channel coding is to encode the Formally, a channel is described by a triple C = (Σ, Ω, p), where information sent over a communication channel in such a Σ is an input alphabet way that in the presence of some channel noise, errors can Ω is an output alphabet p is a probability distribution on Σ Ω and for i Σ, o Ω, p(i, o) is the be detected and/or corrected. × ∈ ∈ probability that the output of the channel is o if the input is i. IMPORTANT CHANNELS There are two basic coding methods Binary symmetric channel maps, with probability p0 each binary input into the BEC (BackwardErrorCorrection) Coding allows the oposite one. Therefore, Pr(0, 1) = Pr(1, 0) = p0 and Pr(0, 0) = Pr(1, 1) = 1 p0. − Binary erasure channel with probability p0, maps binary inputs into outputs in receiver only errors detection. If an error is 0, 1, e , where e is so called the erasure symbol, and Pr(0, 0) = Pr(1, 1) = Pr(p0), { } detected the sender is requested to retransmit Pr(0, e) = Pr(1, e) = 1 p0. − White noise Gaussian channel that models errors in deep space. the message. FEC(ForwardErrorCorrection) Coding allows the receiver to correct a certain amount of errors. prof. Jozef Gruska IV054 1. Basics of coding theory 5/51 prof. Jozef Gruska IV054 1. Basics of coding theory 6/51 WHY WE NEED TO IMPROVE ERROR-CORRECTING CODES BASIC IDEA of ERROR CORRECTION When error correcting capabilities of some code are improved - that is a better code is found -this has the following impacts: Details of the techniques used to protect information against noise in practice are sometimes rather complicated, but basic principles are easily understood. For the same quality of of received information it is possible to achieve that the transmission system operates in more severe conditions; For example; The key idea is that in order to protect a message against a noise, we should encode the message by adding some redundant information to the message. 1 It is possible to reduce the size of antennas or solar panels and the weight of batteries; 2 In space systems such savings can be measured in hundred of thousands of dollars; 3 In mobile telephone systems, improving the code enables the operators to increase the In such a case, even if the message is corrupted by a noise, there will be enough potential number of users in each cell. redundancy in the encoded message to recover – to decode the message Another field of applications of error-correcting codes is that of mass memories: completely. computer hard drives, CD-ROMs, DVDs and so on prof. Jozef Gruska IV054 1. Basics of coding theory 7/51 prof. Jozef Gruska IV054 1. Basics of coding theory 8/51 MAJORITY VOTING DECODING - BASIC IDEA EXAMPLE The basic idea of so called majority voting decoding/principle or of maximal likelihood In case the encoding 0 000 1 111, decoding/principle is → → is used, then the probability of the bit error is p 1 , and the following majority voting ≤ 2 decoding 000, 001, 010, 100 000 and 111, 110, 101, 011 111 to decode a received message w 0 → → is used, then the probability of an erroneous decoding (for the case of 2 or 3 errors) is 3p2(1 p) + p3 = 3p2 2p3 < p by a codeword w that is the closest one to w 0 − − in the whole set of the potential codewords of a given code C. prof. Jozef Gruska IV054 1. Basics of coding theory 9/51 prof. Jozef Gruska IV054 1. Basics of coding theory 10/51 EXAMPLE: Coding of a path avoiding an enemy territory BASIC TERMINOLOGY Datawords - words of a message Story Alice and Bob share an identical map (Fig. 1) gridded as shown in Fig.1. Only Codewords - words of some code. Alice knows the route through which Bob can reach her avoiding the enemy territory. Block code - a code with all codewords of the same length. Alice wants to send Bob the following information about the safe route he should take. Alice Basic assumptions about channels NNWNNWWSSWWNNNNWWN HQ 1 Code length preservation Each output word of a channel has the same length as Three ways to encode the safe route from Bob to the input codeword. Alice are: N 2 Independence of errors The probability of any one symbol being affected in 1 C1 = N = 00, W = 01, S = 11, E = 10 transmissions is the same. { } In such a case any error in the code word Basic strategy for decoding 000001000001011111010100000000010100 For decoding we use the so-called maximal likehood principle, or nearest neighbor would be a disaster. x Bob decoding strategy, or majority voting decoding strategy which says that 2 C2 = 000, 011, 101, 110 Fig. 1 { } A single error in encoding each of symbols N, W, S, E can be detected. the receiver should decode a received word w’ 3 C3 = 00000, 01101, 10110, 11011 { } A single error in decoding each of symbols N, W, S, E can be corrected. as the codewordw that is the closest one to w’. prof. Jozef Gruska IV054 1. Basics of coding theory 11/51 prof. Jozef Gruska IV054 1. Basics of coding theory 12/51 HAMMING DISTANCE BINARY SYMMETRIC CHANNEL The intuitive concept of “closeness“ of two words is well formalized through Hamming Consider a transition of binary symbols such that each symbol has probability of error distance h(x, y) of words x, y. For two words x, y 1 p < 2 . h(x, y) = the number of symbols in which the words x and y differ. 01 - p 0 Example: h(10101, 01100) = 3, h(fourth, eighth) = 4 p Binary symmetric channel Properties of Hamming distance p 1 - p 11 1 h(x, y) = 0 x = y ⇔ If n symbols are transmitted, then the probability of t errors is 2 h(x, y) = h(y, x) t n t n 3 h(x, z) h(x, y) + h(y, z) triangle inequality p (1 p) − ≤ − t An important parameter of codes C is their minimal distance. In the case of binary symmetric channels, the ”nearest` ´ neighbour decoding strategy” is h(C) = min h(x, y) x, y C, x = y , { | ∈ 6 } also ”maximum likelihood decoding strategy”. Therefore, h(C) is the smallest number of errors that can change one codeword into Example Consider C = 000, 111 and the nearest neighbour decoding strategy. { } another. Probability that the received word is decoded correctly Basic error correcting theorem as 000 is (1 p)3 + 3p(1 p)2, − 3 − 2 1 A code C can detect up to s errors if h(C) s + 1. as 111 is (1 p) + 3p(1 p) , ≥ − − 2 A code C can correct up to t errors if h(C) 2t + 1. 3 2 ≥ Therefore Perr (C) = 1 ((1 p) + 3p(1 p) ) Proof (1) Trivial.
Recommended publications
  • An Introduction to Information Theory
    An Introduction to Information Theory Vahid Meghdadi reference : Elements of Information Theory by Cover and Thomas September 2007 Contents 1 Entropy 2 2 Joint and conditional entropy 4 3 Mutual information 5 4 Data Compression or Source Coding 6 5 Channel capacity 8 5.1 examples . 9 5.1.1 Noiseless binary channel . 9 5.1.2 Binary symmetric channel . 9 5.1.3 Binary erasure channel . 10 5.1.4 Two fold channel . 11 6 Differential entropy 12 6.1 Relation between differential and discrete entropy . 13 6.2 joint and conditional entropy . 13 6.3 Some properties . 14 7 The Gaussian channel 15 7.1 Capacity of Gaussian channel . 15 7.2 Band limited channel . 16 7.3 Parallel Gaussian channel . 18 8 Capacity of SIMO channel 19 9 Exercise (to be completed) 22 1 1 Entropy Entropy is a measure of uncertainty of a random variable. The uncertainty or the amount of information containing in a message (or in a particular realization of a random variable) is defined as the inverse of the logarithm of its probabil- ity: log(1=PX (x)). So, less likely outcome carries more information. Let X be a discrete random variable with alphabet X and probability mass function PX (x) = PrfX = xg, x 2 X . For convenience PX (x) will be denoted by p(x). The entropy of X is defined as follows: Definition 1. The entropy H(X) of a discrete random variable is defined by 1 H(X) = E log p(x) X 1 = p(x) log (1) p(x) x2X Entropy indicates the average information contained in X.
    [Show full text]
  • L11-IT-Handouts.Pdf
    Today’s Topics Information Theory • Communication Channel • Noiseless binary channel Mohamed Hamada • Binary Symmetric Channel (BSC) Software Engineering Lab The University of Aizu • Symmetric Channel Email: [email protected] URL: http://www.u-aizu.ac.jp/~hamada • Mutual Information • Channel Capacity 1 Digital Communication Systems Digital Communication Systems 1. Huffman Code. 1. Memoryless 2. Two-pass Huffman Code. 2. Stochastic 3. Lemple-Ziv Code. 3. Markov 4. Fano code. Information User of Information 4. Ergodic User of Source 5. Shannon Code. Information Source Information 6. Arithmetic Code. Source Source Source Source Encoder Decoder Encoder Decoder Channel Channel Channel Channel Encoder Decoder Encoder Decoder Modulator De-Modulator Modulator De-Modulator Channel Channel 2 3 INFORMATION TRANSFER ACROSS CHANNELS Communication Channel Sent Received A (discrete ) channel is a system consisting of an input alphabet messages messages X and output alphabet Y and a probability transition matrix symbols p(y|x) that expresses the probability of observing the output symbol y given that we send the symbol x Channel Channel Source Source Channel sourcecoding coding decoding decoding receiver Examples of channels: Compression Error Correction Decompression CDs, CD – ROMs, DVDs, phones, Source Entropy Channel Capacity Ethernet, Video cassettes etc. Rate vs Distortion Capacity vs Efficiency 4 5 1 Communication Channel Noiseless binary channel Noiseless binary channel Channel Channel 00 input x p(y|x) output y Transition probabilities 11 Memoryless: - output only on input Transition Matrix - input and output alphabet finite 01 p(y | x) = 0 10 1 01 6 7 Binary Symmetric Channel (BSC) Binary Symmetric Channel (BSC) (Noisy channel) (Noisy channel) 00BSC Channel 1 p 1-p 1-p Error Source 00 e BSC Channel p 110 y = x e p 1-p xi i i + 11 Input Output p 1-p 00 1-p 1 1 p BSC Channel 8 9 Symmetric Channel (Noisy channel) Channel XY In the transmission matrix of this channel , all the rows are permutations of each other and so the columns.
    [Show full text]
  • Networked Distributed Source Coding
    Networked distributed source coding Shizheng Li and Aditya Ramamoorthy Abstract The data sensed by differentsensors in a sensor network is typically corre- lated. A natural question is whether the data correlation can be exploited in innova- tive ways along with network information transfer techniques to design efficient and distributed schemes for the operation of such networks. This necessarily involves a coupling between the issues of compression and networked data transmission, that have usually been considered separately. In this work we review the basics of classi- cal distributed source coding and discuss some practical code design techniques for it. We argue that the network introduces several new dimensions to the problem of distributed source coding. The compression rates and the network information flow constrain each other in intricate ways. In particular, we show that network coding is often required for optimally combining distributed source coding and network information transfer and discuss the associated issues in detail. We also examine the problem of resource allocation in the context of distributed source coding over networks. 1 Introduction There are various instances of problems where correlated sources need to be trans- mitted over networks, e.g., a large scale sensor network deployed for temperature or humidity monitoring over a large field or for habitat monitoring in a jungle. This is an example of a network information transfer problem with correlated sources. A natural question is whether the data correlation can be exploited in innovative ways along with network information transfer techniques to design efficient and distributed schemes for the operation of such networks. This necessarily involves a Shizheng Li · Aditya Ramamoorthy Iowa State University, Ames, IA, U.S.A.
    [Show full text]
  • Week 8 Information Channels Suppose That Data Must Be Sent Through a Channel Before It Can Be Used
    CSC 310: Information Theory University of Toronto, Fall 2011 Instructor: Radford M. Neal Week 8 Information Channels Suppose that data must be sent through a channel before it can be used. This channel may be unreliable, but we can use a code designed counteract this. Some questions we aim to answer: • Can we quantify how much information a channel can transmit? • If we have low tolerance for errors, will we be able to make full use of a channel, or must some of the channel’s capacity be lost to ensure a low error probability? • How can we correct (or at least detect) errors in practice? • Can we do as well in practice as the theory says is possible? Error Correction for Memory Blocks The “channel” may transmit information through time rather than space — ie, it is a memory device. Many memory devices store data in blocks — eg, 64 bits for RAM, 512 bytes for disk. Can we correct some errors by adding a few more bits? For instance, could we correct any single error if we use 71 bits to encode a 64 bit block of data stored in RAM? Error Correction in a Communications System In other applications, data arrives in a continuous stream. An overall system might look like this: Encoder for Data often sent in blocks Data Compression Error-Correcting In Program Code Noise Channel Decoder for Decompression Data Error-Correcting Code Program Out Error Detection We might also be interested in detecting errors, even if we can’t correct them: • For RAM or disk memory, error detection tells us that we need to call the repair person.
    [Show full text]
  • A Survey of Results for Deletion Channels and Related Synchronization Channels∗
    Probability Surveys Vol. 6 (2009) 1–33 ISSN: 1549-5787 DOI: 10.1214/08-PS141 A survey of results for deletion channels and related synchronization channels∗ Michael Mitzenmacher† Harvard University, School of Engineering and Applied Sciences e-mail: [email protected] Abstract: The purpose of this survey is to describe recent progress in the study of the binary deletion channel and related channelswith synchroniza- tion errors, including a clear description of open problems in this area, with the hope of spurring further research. As an example, while the capacity of the binary symmetric error channel and the binary erasure channel have been known since Shannon, we still do not have a closed-form description of the capacity of the binary deletion channel. We highlight a recent result that shows that the capacity is at least (1 − p)/9 when each bit is deleted independently with fixed probability p. AMS 2000 subject classifications: Primary 94B50; secondary 68P30. Keywords and phrases: Deletion channels, synchronization channels, capacity bounds, random subsequences. Received November 2008. Contents 1 Introduction................................. 2 2 The ultimate goal: A maximum likelihoodargument . 4 3 Shannon-stylearguments . 6 3.1 Howthebasicargumentfails . 7 3.2 Codebooks fromfirstorder Markovchains . 8 4 Better decoding via jigsaw puzzles . 12 5 A useful reduction for deletion channels . 16 6 A related information theoretic formulation . .... 17 7 Upper bounds on the capacity of deletion channels . ... 19 8 Variationsonthedeletionchannel . 22 8.1 Stickychannels ............................ 22 8.2 Segmented deletion and insertion channels . 24 8.3 Deletionsoverlargeralphabets . 26 9 Tracereconstruction ... .... ... .... ... .... ... .... 27 10Conclusion ................................. 29 References.................................... 30 ∗This is an original survey paper.
    [Show full text]
  • Lecture Notes on Channel Coding
    Lecture Notes on Channel Coding Georg B¨ocherer Institute for Communications Engineering Technical University of Munich, Germany [email protected] July 5, 2016 arXiv:1607.00974v1 [cs.IT] 4 Jul 2016 These lecture notes on channel coding were developed for a one-semester course for graduate students of electrical engineering. Chapter 1 reviews the basic problem of channel coding. Chapters 2{5 are on linear block codes, cyclic codes, Reed-Solomon codes, and BCH codes, respectively. The notes are self-contained and were written with the intent to derive the presented results with mathematical rigor. The notes contain in total 68 homework problems, of which 20% require computer programming. Contents 1 Channel Coding7 1.1 Channel . .7 1.2 Encoder . .8 1.3 Decoder . .8 1.3.1 Observe the Output, Guess the Input . .8 1.3.2 MAP Rule . .9 1.3.3 ML Rule . 10 1.4 Block Codes . 10 1.4.1 Probability of Error vs Transmitted Information . 10 1.4.2 Probability of Error, Information Rate, Block Length . 12 1.4.3 ML Decoder . 14 1.5 Problems . 16 2 Linear Block Codes 20 2.1 Basic Properties . 20 2.1.1 Groups and Fields . 20 2.1.2 Vector Spaces . 22 2.1.3 Linear Block Codes . 23 2.1.4 Generator Matrix . 23 2.2 Code Performance . 25 2.2.1 Hamming Geometry . 25 2.2.2 Bhattacharyya Parameter . 27 2.2.3 Bound on Probability of Error . 27 2.3 Syndrome Decoding . 29 2.3.1 Dual Code . 31 2.3.2 Check Matrix .
    [Show full text]
  • Lecture 21: Computing Capacities, Coding in Practice, & Review
    COMP2610/6261 - Information Theory Lecture 21: Computing Capacities, Coding in Practice, & Review ANU Logo UseMark Guidelines Reid and Aditya Menon Research School of Computer Science The ANU logo is a contemporary The Australian National University refection of our heritage. It clearly presents our name, our shield and our motto: First to learn the nature of things. To preserve the authenticity of our brand identity, there are rules that govern how our logo is used. Preferred logo Black version Preferred logo - horizontal logo October 14, 2013 The preferred logo should be used on a white background. This version includes black text with the crest in Deep Gold in either PMS or CMYK. Black Where colour printing is not available, the black logo can be used on a white background. Reverse Mark Reid and Aditya Menon (ANU) COMP2610/6261 - Information Theory Oct. 14, 2014 1 / 21 The logo can be used white reversed out of a black background, or occasionally a neutral dark background. Deep Gold Black C30 M50 Y70 K40 C0 M0 Y0 K100 PMS Metallic 8620 PMS Process Black PMS 463 Reverse version Any application of the ANU logo on a coloured background is subject to approval by the Marketing Offce, contact [email protected] LOGO USE GUIDELINES 1 THE AUSTRALIAN NATIONAL UNIVERSITY 1 Computing Capacities 2 Good Codes vs. Practical Codes 3 Linear Codes 4 Coding: Review Mark Reid and Aditya Menon (ANU) COMP2610/6261 - Information Theory Oct. 14, 2014 2 / 21 1 Computing Capacities 2 Good Codes vs. Practical Codes 3 Linear Codes 4 Coding: Review Mark Reid and Aditya Menon (ANU) COMP2610/6261 - Information Theory Oct.
    [Show full text]
  • Noisy-Channel Coding Copyright Cambridge University Press 2003
    Copyright Cambridge University Press 2003. On-screen viewing permitted. Printing not permitted. http://www.cambridge.org/0521642981 You can buy this book for 30 pounds or $50. See http://www.inference.phy.cam.ac.uk/mackay/itila/ for links. Part II Noisy-Channel Coding Copyright Cambridge University Press 2003. On-screen viewing permitted. Printing not permitted. http://www.cambridge.org/0521642981 You can buy this book for 30 pounds or $50. See http://www.inference.phy.cam.ac.uk/mackay/itila/ for links. 8 Dependent Random Variables In the last three chapters on data compression we concentrated on random vectors x coming from an extremely simple probability distribution, namely the separable distribution in which each component xn is independent of the others. In this chapter, we consider joint ensembles in which the random variables are dependent. This material has two motivations. First, data from the real world have interesting correlations, so to do data compression well, we need to know how to work with models that include dependences. Second, a noisy channel with input x and output y defines a joint ensemble in which x and y are dependent { if they were independent, it would be impossible to communicate over the channel { so communication over noisy channels (the topic of chapters 9{11) is described in terms of the entropy of joint ensembles. 8.1 More about entropy This section gives definitions and exercises to do with entropy, carrying on from section 2.4. The joint entropy of X; Y is: 1 H(X; Y ) = P (x; y) log : (8.1) P (x; y) xy2AXX AY Entropy is additive for independent random variables: H(X; Y ) = H(X) + H(Y ) iff P (x; y) = P (x)P (y): (8.2) The conditional entropy of X given y = bk is the entropy of the proba- bility distribution P (x y = b ).
    [Show full text]
  • Channel Capacity, Rate of Channel Code 15.1 the Information Theory
    10-704: Information Processing and Learning Spring 2012 Lecture 15: Channel Capacity, Rate of Channel Code Lecturer: Aarti Singh Scribes: Martin Azizyan Disclaimer: These notes have not been subjected to the usual scrutiny reserved for formal publications. They may be distributed outside this class only with the permission of the Instructor. 15.1 The Information Theory Diagram Until now we have studied the data compression (source coding) problem, schematically described by the following diagram: Source Compressor Decompressor Receiver We now turn to the channel coding problem, where we introduce an intermediate step between the output of the compressor and the input to the decompressor: Source Compressor Channel Decompressor Receiver A \channel" is any source of noise, for example a wireless link, an error prone storage device, an IP network with the potential for packet loss, etc. In the presence of such noise we need to introduce some redundancy in the signal to ensure successful transmission, so we add two new steps in the diagram: Compressed Input X Output Y To Encoder Channel Decoder Message Decompressor Essentially, a channel is a conditional distribution p(Y jX). We will focus on discrete memoryless chan- nels (DMC): Discrete: Input and output alphabets X ; Y are finite; and Memoryless: p(Y jX) is independent of outputs in previous timesteps (later we will discuss more general feedback channels). Definition 15.1 The information capacity of a (DMC) channel is C = max I(X; Y ) p(x) where the maximum is over all distributions on X. 15-1 15-2 Lecture 15: Channel Capacity, Rate of Channel Code Informally, the operational capacity of a channel is the highest rate in terms of bits/channel use (e.g.
    [Show full text]
  • Channel Capacity and Coding
    ECE 515 Information Theory Channel Capacity and Coding 1 Information Theory Problems • How to transmit or store information as efficiently as possible. • What is the maximum amount of information that can be transmitted or stored reliably? • How can information be kept secure? 2 Digital Communications System W X PX Data source PY|X Data sink Ŵ Y PY Ŵ is an estimate of W 3 Communication Channel • Cover and Thomas Chapter 7 4 Discrete Memoryless Channel X Y x1 y1 x2 y2 . xK yJ J≥K≥2 5 Discrete Memoryless Channel Channel Transition Matrix py(|)11 x py (|)1 xkK py (|)1 x P= py(|)j x py (|)jk x py (|)jK x 1 py(|)J x1 py (|)Jk x py (|)JK x 6 Binary Symmetric Channel x1 y1 x2 y2 channel matrix pp pp 7 Binary Errors with Erasure Channel 1-p-q x 0 0 y 1 q 1 p e y2 p q x2 1 1 y3 1-p-q 8 BPSK Modulation K=2 2E x t = + b cos 2π ft 0 bit 1 ( ) ( c ) Eb is the energy per bit Tb Tb is the bit duration 2E f is the carrier x t = − b cos 2π ft 1 bit c 2 ( ) ( c ) frequency Tb 9 BPSK Demodulation in AWGN − Eb 0 + Eb x2 x1 p(y|x2) p(y|x1) 0 x2 x1 2 1 (yx− k ) pYX (yx |= xk ) = exp− | 2 2 2πσ 2σ 10 BPSK Demodulation K=2 − + Eb 0 Eb J=3 y3 y2 y1 − + Eb 0 Eb J=4 y4 y3 y2 y1 − + Eb 0 Eb J=8 y8 y1 11 Mutual Information for a BSC crossover probability p X BSC Y pp=1 − channel matrix pp PY|X = pp X Y p w 0 0 p p(xw= 0) = p p(x==−= 1) 1 ww w 1 1 p 12 13 Convex Functions Concave Convex 14 Concave Function 15 Convex Function 16 Mutual Information 17 18 input probabilities 19 20 21 channel transition probabilities 22 BSC I(X;Y) w p 23 Channel Capacity X channel
    [Show full text]
  • ECE 587 / STA 563: Lecture 6 – Channel Coding Information Theory Duke University, Fall 2020
    ECE 587 / STA 563: Lecture 6 { Channel Coding Information Theory Duke University, Fall 2020 Author: Galen Reeves Last Modified: September 24, 2020 Outline of lecture: 6.1 Introduction to Channel Coding.............................1 6.1.1 Motivation.....................................1 6.1.2 Problem Setup...................................3 6.1.3 Channel Coding Theorem.............................5 6.2 Examples of Discrete Memoryless Channels.......................6 6.2.1 Binary Erasure Channel (BEC)..........................6 6.2.2 Binary Symmetric Channel (BSC)........................7 6.3 Converse to the Channel Coding Theorem:.......................7 6.4 Proof of Channel Coding Theorem via Random Coding................9 6.4.1 Encoding & Decoding............................... 10 6.4.2 Performance Analysis............................... 11 6.4.3 Strengthening the Proof.............................. 13 6.5 Channel Coding with Feedback.............................. 15 6.6 Coding Theory....................................... 16 6.6.1 Hamming Codes.................................. 16 6.6.2 Beyond Hamming................................. 17 6.6.3 The Hat Game................................... 18 6.1 Introduction to Channel Coding 6.1.1 Motivation • Communication problems play an important role in many applications: ◦ Wireless communication between your laptop and a router ◦ Wireless communication between your cell phone and a cell tower (e.g, GSM and CDMA) ◦ Fiber-optic communication across the internet ◦ Data storage on a hard drive (communicating from the past to the future). ◦ Communication between Hubble Space Telescope and Earth. ◦ Statistics: communicating the true state of nature θ through the observed data X • The primary goals of communication are: (1) Reliability: We want the received message to be equal to the transmitted message (at least with very high probability) Error probability: P[received message 6= transmitted message] 2 ECE 587 / STA 563: Lecture 6 (2) Efficiency: We want to communicate each message as quickly as possible.
    [Show full text]
  • Channel Capacity
    Chapter 7 Channel Capacity In Section 5.2 the block diagram of a communication system in Figure 5.1 was discussed. It was stated that the source coding and the channel coding can be separated. In Chap- ters 5 and 6 the source coding was treated. Here we will concentrate on the analysis of the channel coding system. In Figure 7.1 a block diagram is shown for the channel encoder, the channel and the channel decoder. The channel in this figure is a mathematical model that represents ev- erything that can occur in the actual transmission. This includes e.g. background noise, scratches on the surface of a CD or erasures due to overflow in router buffers. U Channel X Y Channel Ub Channel encoder decoder Figure 7.1: A model of the channel coding system. A little more formal description of the encoding scheme is as follows. • The symbols U are assumed to be taken from an index set I = f1; 2;:::;Mg. • The encoding function x : I!X is a mapping from the index set I to the set of codewords X . We can denote the codewords as x(1); x(2); : : : ; x(M). In the most cases in this text, we will assume that the codewords are binary vectors of length k, X 2 f0; 1gn. • As the codeword is transmitted over the channel, errors occur and the received vectors are y 2 Y. In many situations the received symbols are taken from a larger alphabet than the code symbols. It can be that Y are real values detected by the receiver, while the code symbols X are binary.
    [Show full text]