Distance Path Because, 01 1 Regardless of What Happens Subsequently, This Path Will 0 × 4 Have a Larger Hamming 11 1/10 (1) 1 Distance from Y

Distance Path Because, 01 1 Regardless of What Happens Subsequently, This Path Will 0 × 4 Have a Larger Hamming 11 1/10 (1) 1 Distance from Y

Direct Minimum Distance Decoding y Suppose = [11 01 11]. y Find . y Find the message which corresponds to the (valid) codeword with minimum (Hamming) distance from . y ܠപൌො ݀ ܠപǡܡ ܠപ പ + + 89 Direct Minimum Distance Decoding + y Suppose = [11 01 11]. y Find . + y Find the message which corresponds to the (valid) codeword with minimum (Hamming) distance from . ࢟ = [ 11 01 11 ]. 00 0/00 0/00 0/00 For 3-bit message, 3 10 there are 2 = 8 possible codewords. We can list all possible codewords. However, here, let’s first try to work 01 on the distance directly. 11 1/10 90 Direct Minimum Distance Decoding y Suppose = [11 01 11]. y Find . y Find the message which corresponds to the (valid) codeword with minimum (Hamming) distance from . The number in parentheses on each ࢟ = [ 11 01 11 ]. branch is the branch metric, obtained 0/00 (2) 0/00 (1) 0/00 (2) by counting the differences between 00 the encoded bits and the corresponding bits in ࢟. 10 01 11 1/10 (1) 91 Direct Minimum Distance Decoding y Suppose = [11 01 11]. y Find . y Find the message which corresponds to the (valid) codeword with minimum (Hamming) distance from . ࢟ = [ 11 01 11 ]. b d(x,y) 0/00 (2) 0/00 (1) 0/00 (2) 00 000 2+1+2 = 5 001 2+1+0 = 3 010 2+1+1 = 4 10 011 2+1+1 = 4 100 0+2+0 = 2 01 101 0+2+2 = 4 110 0+0+1 = 1 11 1/10 (1) 111 0+0+1 = 1 92 Viterbi decoding y Developed by Andrew J. Viterbi y Also co-founded Qualcomm Inc. y Published in the paper “Error Bounds for Convolutional Codes and an Asymptotically Optimum Decoding Algorithm”, IEEE Transactions on Information Theory, Volume IT-13, pages 260-269, in April, 1967. https://en.wikipedia.org/wiki/Andrew_Viterbi 93 Viterbi and His Decoding Algorithm 94 [http://viterbi.usc.edu/about/viterbi/viterbi_video.htm] Andrew J. Viterbi y 1991: Claude E. Shannon Award y 1952-1957: MIT BS & MS y Studied electronics and communications theory under such renowned scholars as Norbert Wiener, Claude Shannon, Bruno Rossi and Roberto Fano. y 1962: Earned one of the first doctorates in electrical engineering granted at the University of Southern California (USC) y Ph.D. dissertation: error correcting codes y 2004: USC Viterbi School of Engineering 95 named in recognition of his $52 million gift Andrew J. Viterbi 96 [http://viterbi.usc.edu/about/viterbi/viterbi_video.htm] Andrew J. Viterbi y Cofounded Qualcomm y Helped to develop the CDMA standard for cellular networks. y 1998 Golden Jubilee Award for Technological Innovation y To commemorate the 50th Anniversary of Information Theory y Given to the authors of discoveries, advances and inventions that have had a profound impact in the technology of information transmission, processing and compression. 1. Norman Abramson: For the invention of the first random-access communication protocol. 2. Elwyn Berlekamp: For the invention of a computationally efficient algebraic decoding algorithm. 3. Claude Berrou, Alain Glavieux and Punya Thitimajshima: For the invention of turbo codes. 4. Ingrid Daubechies: For the invention of wavelet-based methods for signal processing. 5. Whitfield Diffie and Martin Hellman: For the invention of public-key cryptography. 6. Peter Elias: For the invention of convolutional codes. 7. G. David Forney, Jr: For the invention of concatenated codes and a generalized minimum-distance decoding algorithm. 8. Robert M. Gray: For the invention and development of training mode vector quantization. 9. David Huffman: For the invention of the Huffman minimum-length lossless data- compression code. 10. Kees A. Schouhamer Immink: For the invention of constrained codes for commercial recording systems. 11. Abraham Lempel and Jacob Ziv: For the invention of the Lempel-Ziv universal data compression algorithm. 12. Robert W. Lucky: For the invention of pioneering adaptive equalization methods. 13. Dwight O. North: For the invention of the matched filter. 14. Irving S. Reed: For the co-invention of the Reed-Solomon error correction codes. 15. Jorma Rissanen: For the invention of arithmetic coding. 16. Gottfried Ungerboeck: For the invention of trellis coded modulation. 17.Andrew J. Viterbi: For the invention of the Viterbi algorithm. 97 Viterbi Decoding: Ex. 1 y Suppose = [11 01 11]. y Find . y Find the message which corresponds to the (valid) codeword with minimum (Hamming) distance from . ࢟ = [ 11 01 11 ]. 2 0/00 (2) 0/00 (1) 30/00 (2) 00 Each circled number at a node is the running 0 3 10 (cumulative) path metric, obtained by summing 2 01 branch metrics (distance) up to that node. Here, it is 0 11 1/10 (1) simply the cumulative 98 distance. Viterbi Decoding: Ex. 1 y Suppose = [11 01 11]. y Find . y Find the message which corresponds to the (valid) codeword with minimum (Hamming) distance from . y For the last column of nodes, ࢟ = [ 11 01 11 ]. 2 each of the nodes has two 0/00 (2) 0/00 (1) 30/00 (2) 00 branches going into it. y So, there are two possible 0 3 cumulative distance values. 10 y We discard the larger- 2 metric path because, 01 regardless of what happens subsequently, this path will 0 have a larger Hamming 11 1/10 (1) distance from y . 99 Viterbi Decoding: Ex. 1 y Suppose = [11 01 11]. y Find . y Find the message which corresponds to the (valid) codeword with minimum (Hamming) distance from . ࢟ = [ 11 01 11 ]. 2 3 5 0/00 (2) 0/00 (1) 0/00 (2) min(3+2,2+0) 00 2 We discard the larger-distance 0 3 3 path because, 10 min(3+0,2+2) 4 regardless of what happens subsequently, 2 4 min(3+1,0+1) this path will have a 01 1 larger Hamming 0 4 distance from y. 11 min(3+1,0+1) 1/10 (1) 1 100 Viterbi Decoding: Ex. 1 y Suppose = [11 01 11]. y Find . y Find the message which corresponds to the (valid) codeword with minimum (Hamming) distance from . y For the last column of nodes, ࢟ = [ 11 01 11 ]. each of the nodes has two 2 3 5 0/00 (2) 0/00 (1) 0/00 (2) branches going into it. 00 × 2 y So, there are two possible 0 3 3 cumulative distance values. 10 × 4 y We discard the larger- 2 × 4 distance path because, 01 1 regardless of what happens subsequently, this path will 0 × 4 have a larger Hamming 11 1/10 (1) 1 distance from y. 101 Viterbi Decoding: Ex. 1 y Suppose = [11 01 11]. y Find . y Find the message which corresponds to the (valid) codeword with minimum (Hamming) distance from . y For the last column of nodes, ࢟ = [ 11 01 11 ]. 2 each of the nodes has two 0/00 (2) 0/00 (1) 3 2 00 branches going into it. y So, there are two possible 0 3 3 cumulative distance values. 10 y We discard the larger- 2 1 distance path because, 01 regardless of what happens subsequently, this path will 0 1 have a larger Hamming 11 1/10 (1) distance from y . 102 Viterbi Decoding: Ex. 1 y Suppose = [11 01 11]. y Find . y Find the message which corresponds to the (valid) codeword with minimum (Hamming) distance from . ࢟ = [ 11 01 11 ]. 2 0/00 (2) 0/00 (1) 3 2 00 Note that we keep exactly one 0 3 3 10 (optimal) survivor path to each state. 2 1 01 (Unless there is a tie, then we keep both 0 1 or choose any.) 11 1/10 (1) 103 Viterbi Decoding: Ex. 1 y Suppose = [11 01 11]. y Find . y Find the message which corresponds to the (valid) codeword with minimum (Hamming) distance from . ࢟ = [ 11 01 11 ]. y So, the codewords 2 0/00 (2) 0/00 (1) 3 2 00 which are nearest to is [11 01 01] or [11 01 0 3 3 10 10]. y 2 1 The corresponding 01 messages are [110] or 0 1 [111], respectively. 11 1/10 (1) 104 Viterbi Decoding: Ex. 2 y Suppose = [11 01 11 00 01 10]. y Find . same as before ࢟ = [ 11 01 11 00 01 10 ]. 2 0/00 (2) 0/00 (1) 3 2 00 The first part is the same as 0 3 3 10 before. So, we simply copy the diagram that we had. 2 1 01 0 1 11 1/10 (1) 105 Viterbi Decoding: Ex. 2 y Suppose = [11 01 11 00 01 10]. y Find . same as before ࢟ = [ 11 01 11 00 01 10 ]. 2 0/00 (2) 0/00 (1) 3 2 0/00 (0) 00 min(2+0,1+2) 0 3 3 10 min(2+2,1+0) 2 1 01 min(3+1,1+1) 0 1 min(3+1,1+1) 11 1/10 (1) 1/10 (1) 106 Viterbi Decoding: Ex. 2 y Suppose = [11 01 11 00 01 10]. y Find . ࢟ = [ 11 01 11 00 01 10 ]. 2 0/00 (2) 0/00 (1) 3 2 0/00 (0) 00 min(2+0,1+2) = 2 × 0 3 3 10 × min(2+2,1+0) = 1 2 1 01 × min(3+1,1+1) = 2 0 1 × min(3+1,1+1) = 2 11 1/10 (1) 1/10 (1) 107 Viterbi Decoding: Ex. 2 y Suppose = [11 01 11 00 01 10]. y Find . ࢟ = [ 11 01 11 00 01 10 ]. 2 0/00 (2) 0/00 (1) 3 2 0/00 (0) 2 0/00 (1) 00 min(2+1,2+1) = 3 0 3 3 1 10 min(2+1,2+1) = 3 2 1 2 01 min(1+2,2+0) = 2 0 1 2 min(1+0,2+2) = 1 11 1/10 (1) 1/10 (1) 1/10 (2) 108 Viterbi Decoding: Ex.

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    12 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us