Andrew J. and Erna Viterbi Family Archives, 1905-20070335

Total Page:16

File Type:pdf, Size:1020Kb

Andrew J. and Erna Viterbi Family Archives, 1905-20070335 http://oac.cdlib.org/findaid/ark:/13030/kt7199r7h1 Online items available Finding Aid for the Andrew J. and Erna Viterbi Family Archives, 1905-20070335 A Guide to the Collection Finding aid prepared by Michael Hooks, Viterbi Family Archivist The Andrew and Erna Viterbi School of Engineering, University of Southern California (USC) First Edition USC Libraries Special Collections Doheny Memorial Library 206 3550 Trousdale Parkway Los Angeles, California, 90089-0189 213-740-5900 [email protected] 2008 University Archives of the University of Southern California Finding Aid for the Andrew J. and Erna 0335 1 Viterbi Family Archives, 1905-20070335 Title: Andrew J. and Erna Viterbi Family Archives Date (inclusive): 1905-2007 Collection number: 0335 creator: Viterbi, Erna Finci creator: Viterbi, Andrew J. Physical Description: 20.0 Linear feet47 document cases, 1 small box, 1 oversize box35000 digital objects Location: University Archives row A Contributing Institution: USC Libraries Special Collections Doheny Memorial Library 206 3550 Trousdale Parkway Los Angeles, California, 90089-0189 Language of Material: English Language of Material: The bulk of the materials are written in English, however other languages are represented as well. These additional languages include Chinese, French, German, Hebrew, Italian, and Japanese. Conditions Governing Access note There are materials within the archives that are marked confidential or proprietary, or that contain information that is obviously confidential. Examples of the latter include letters of references and recommendations for employment, promotions, and awards; nominations for awards and honors; resumes of colleagues of Dr. Viterbi; and grade reports of students in Dr. Viterbi's classes at the University of California, Los Angeles, and the University of California, San Diego. These restricted items were not scanned and, therefore, are not included in the USC Digital Archive. Researchers wishing to see any of the restricted materials should consult with the USC Libraries Special Collections staff. Conditions Governing Use note The archives contains published articles and book chapters authored or co-authored by Dr. Viterbi and others. Researchers are reminded of the copyright restrictions imposed by publishers on reusing their articles and chapters. It is the responsibility of researchers to acquire permission from publishers when reusing such materials. Preferred Citation note [Identification of item], Andrew J. and Erna Viterbi Family Archives, Collection 335, University of Southern California, Los Angeles. Immediate Source of Acquisition note The Andrew J. and Erna Viterbi Family Archives was primarily obtained through gifts from Dr. Andrew J. Viterbi. Additional donations were made by Dr. William C. Lindsey, and Dr. Solomon Golomb, both of the Ming Hsieh Department of Electrical Engineering at the University of Southern California (USC). Other materials were collected and by the Viterbi Family Archivist, Dr. Michael Hooks, who obtained them from the JPL Archives, Dr. Viterbi's professional colleagues, and Web sources. Biographical/Historical note Andrew James Viterbi was born on March 9, 1935 in Bergamo, Lombardy, Italy, the only child of Dr. Achille and Maria Viterbi. In 1939, the Viterbi family immigrated to the United States due to the anti-Semitic laws passed in fascist Italy. They lived first in New York City, and then moved to Boston, Massachusetts, when Andrew was six years old. Andrew attended the public schools in Boston, and graduated from Boston Latin School in 1952. He received his B.S. and M.S. degrees from the Massachusetts Institute of Technology (MIT) in 1957, and the Ph.D. from the University of Southern California (USC) in 1962. Andrew Viterbi currently serves as President of the Viterbi Group, LLC, founded in 2000 in San Diego, California. The Viterbi Group advises and invests in startup companies, predominantly in wireless communications, network infrastructure and imaging. In July 1985, Dr. Viterbi co-founded QUALCOMM Incorporated, a developer and manufacturer of mobile satellite communications and digital wireless telephony, where he served as Vice Chairman until 2000 and as Chief Technical Officer until 1996. Under his leadership, QUALCOMM received international recognition for innovative technology in the areas of digital wireless communication systems and products based on Code Division Multiple Access (CDMA) technologies. Previously in 1968, Dr. Viterbi co-founded LINKABIT Corporation, a digital communications company, where he served as Executive Vice President and later as President in the early 1980s. From 1963 to 1973, Dr. Viterbi served as a Professor at the University of California, Los Angeles, (UCLA) School of Engineering and Applied Science, where he did fundamental work in digital communication theory and wrote numerous research Finding Aid for the Andrew J. and Erna 0335 2 Viterbi Family Archives, 1905-20070335 papers and two books, for which he has received international recognition. He continued teaching on a part-time basis at the University of California, San Diego (UCSD), until 1994, where he is currently Professor Emeritus. Also in 2001 he was invited by Technion, Israel Institute of Technology, to become a Distinguished Visiting Professor of Electrical Engineering and in 2004 he was named to the President's Chair in the Department of Electrical Engineering Systems at USC. From 1957 to 1963, Dr. Viterbi was a member of the Communications Research Section of the Jet Propulsion Laboratory (JPL), an operting division of the California Institute of Technology (Caltech) in Pasadena, California. While at JPL, he was one of the first communication engineers to recognize the potential of and propose digital transmission techniques for space and satellite telecommunication systems. Dr. Viterbi has received numerous awards and recognition for his leadership and substantial contributions to communications theory and its industrial applications over the years. He has received honorary doctorates from universities in the United States, Canada, Italy and Israel, and has been otherwise honored in Japan, Germany, Italy and the United States. He is a Fellow of the IEEE, a Marconi Fellow, and a Member of the U.S. National Academy of Engineering and the U.S. National Academy of Sciences and the American Academy of Arts and Sciences. From 1997 until 2001, he served as a member of the U.S. President's Information Technology Advisory Committee. He is currently a trustee of the University of Southern California, a Board Member of the Scripps Research Institute in La Jolla, California, a trustee of the Mathematical Sciences Research Institute in Berkeley, California, and a member of the California Council on Science and Technology. All four international standards for digital cellular telephony utilize the Viterbi algorithm for interference suppression, as do most digital satellite communication systems, both for business applications and for direct satellite broadcast to the home. Erna Finci Viterbi was born on January 20, 1934, in Sarajevo, Yugoslavia. Her parents were Joseph and Lenka Finci. Andrew and Erna married on June 15, 1958, in Los Angeles, California. They have 3 adult children: one daughter, Audrey M., and two sons, Alan R., and Alex. Scope and Contents note The Andrew J. and Erna Viterbi Family Archives documents the career and professional activities of Dr. Andrew Viterbi, noted researcher, scholar, innovator, and businessman, as well as provide information about the Viterbi and Finci families. The professional papers consist of audio materials, awards, certificates, clippings, correspondence, memoranda, manusciprt materials, patents, photographs, presentations, publications and reprints, research materials, and reports that are useful in following Dr. Viterbi's career and provide insight into his constributions to the field of digital communication. The family materials provide information about the Viterbi and Finci families. The materials include certificates, clippings, correspondence, diplomas, drawings, photographs, publications, and research materials. Separated Materials note Oversize materials, such as newspaper pages, manuscript galleys, and drawings, were separated and rehoused in an oversize box. DVDs and an audio tape were separated and rehoused in a small box. Duplicate materials were removed and retained for returning to Dr. Viterbi. Arrangement note The archives is organized into 18 series, arranged alphabetically: Series 1. Academic Affiliations; Series 2. Academic Files; Series 3. Biographical Information; Series 4. Business Plan; Series 5. Committees and Boards/Professional Affiliations; Series 6. Correspondence; Series 7. Honors, with 3 subseries: Academies, Awards, and Honorary Degrees; Series 8. Mementos; Series 9. News Items; Series 10. Oral History; Series 11. Papers Authored; Series 12. Patents; Series 13. Philanthropic Activities; Series 14. Presentations; Series 15. Prior Art; Series 16. Publications; Series 17. Technical Information; and Series 18. Viterbi Family Materials Accruals note The initial accession consists of 17 banker's boxes packed and brought to the Special Collections Department by Michael Hooks, Viterbi Family Archivist, and Jackie Morin, Processing Archivist, on November 28, 2006. Dr. Viterbi delivered an additional box of materials on February 26, 2007. An additional envelope of materias was received from Dr. Viterbi on June 8, 2007. Lastly, colleagues of Dr. Viterbi donated items at the request
Recommended publications
  • Data Compression: Dictionary-Based Coding 2 / 37 Dictionary-Based Coding Dictionary-Based Coding
    Dictionary-based Coding already coded not yet coded search buffer look-ahead buffer cursor (N symbols) (L symbols) We know the past but cannot control it. We control the future but... Last Lecture Last Lecture: Predictive Lossless Coding Predictive Lossless Coding Simple and effective way to exploit dependencies between neighboring symbols / samples Optimal predictor: Conditional mean (requires storage of large tables) Affine and Linear Prediction Simple structure, low-complex implementation possible Optimal prediction parameters are given by solution of Yule-Walker equations Works very well for real signals (e.g., audio, images, ...) Efficient Lossless Coding for Real-World Signals Affine/linear prediction (often: block-adaptive choice of prediction parameters) Entropy coding of prediction errors (e.g., arithmetic coding) Using marginal pmf often already yields good results Can be improved by using conditional pmfs (with simple conditions) Heiko Schwarz (Freie Universität Berlin) — Data Compression: Dictionary-based Coding 2 / 37 Dictionary-based Coding Dictionary-Based Coding Coding of Text Files Very high amount of dependencies Affine prediction does not work (requires linear dependencies) Higher-order conditional coding should work well, but is way to complex (memory) Alternative: Do not code single characters, but words or phrases Example: English Texts Oxford English Dictionary lists less than 230 000 words (including obsolete words) On average, a word contains about 6 characters Average codeword length per character would be limited by 1
    [Show full text]
  • Bernard M. Oliver Oral History Interview
    http://oac.cdlib.org/findaid/ark:/13030/kt658038wn Online items available Bernard M. Oliver Oral History Interview Daniel Hartwig Stanford University. Libraries.Department of Special Collections and University Archives Stanford, California November 2010 Copyright © 2015 The Board of Trustees of the Leland Stanford Junior University. All rights reserved. Note This encoded finding aid is compliant with Stanford EAD Best Practice Guidelines, Version 1.0. Bernard M. Oliver Oral History SCM0111 1 Interview Overview Call Number: SCM0111 Creator: Oliver, Bernard M., 1916- Title: Bernard M. Oliver oral history interview Dates: 1985-1986 Physical Description: 0.02 Linear feet (1 folder) Summary: Transcript of an interview conducted by Arthur L. Norberg covering Oliver's early life, his education, and work experiences at Bell Laboratories and Hewlett-Packard. Subjects include television research, radar, information theory, organizational climate and objectives at both companies, Hewlett-Packard's associations with Stanford University, and Oliver's association with William Hewlett and David Packard. Language(s): The materials are in English. Repository: Department of Special Collections and University Archives Green Library 557 Escondido Mall Stanford, CA 94305-6064 Email: [email protected] Phone: (650) 725-1022 URL: http://library.stanford.edu/spc Information about Access Reproduction is prohibited. Ownership & Copyright All requests to reproduce, publish, quote from, or otherwise use collection materials must be submitted in writing to the Head of Special Collections and University Archives, Stanford University Libraries, Stanford, California 94304-6064. Consent is given on behalf of Special Collections as the owner of the physical items and is not intended to include or imply permission from the copyright owner.
    [Show full text]
  • Group Testing
    Group Testing Amit Kumar Sinhababu∗ and Vikraman Choudhuryy Department of Computer Science and Engineering, Indian Institute of Technology Kanpur April 21, 2013 1 Motivation Out original motivation in this project was to study \coding theory in data streaming", which has two aspects. • Applications of theory correcting codes to efficiently solve problems in the model of data streaming. • Solving coding theory problems in the model of data streaming. For ex- ample, \Can one recognize a Reed-Solomon codeword in one-pass using only poly-log space?" [1] As we started, we were directed to a related combinatorial problem, \Group testing", which is important on its own, having connections with \Compressed Sensing", \Data Streaming", \Coding Theory", \Expanders", \Derandomiza- tion". This project report surveys some of these interesting connections. 2 Group Testing The group testing problem is to identify the set of \positives" (\defectives", or \infected", or 1) from a large set of population/items, using as few tests as possible. ∗[email protected] [email protected] 1 2.1 Definition There is an unknown stream x 2 f0; 1gn with at most d ones in it. We are allowed to test any subset S of the indices. The answer to the test tells whether xi = 0 for all i 2 S, or not (at least one xi = 1). The objective is to design as few tests as possible (t tests) such that x can be identified as fast as possible. Group testing strategies can be either adaptive or non-adaptive. A group testing algorithm is non-adaptive if all tests must be specified without knowing the outcome of other tests.
    [Show full text]
  • DRASIC Distributed Recurrent Autoencoder for Scalable
    DRASIC: Distributed Recurrent Autoencoder for Scalable Image Compression Enmao Diao∗, Jie Dingy, and Vahid Tarokh∗ ∗Duke University yUniversity of Minnesota-Twin Cities Durham, NC, 27701, USA Minneapolis, MN 55455, USA [email protected] [email protected] [email protected] Abstract We propose a new architecture for distributed image compression from a group of distributed data sources. The work is motivated by practical needs of data-driven codec design, low power con- sumption, robustness, and data privacy. The proposed architecture, which we refer to as Distributed Recurrent Autoencoder for Scalable Image Compression (DRASIC), is able to train distributed encoders and one joint decoder on correlated data sources. Its compression capability is much bet- ter than the method of training codecs separately. Meanwhile, the performance of our distributed system with 10 distributed sources is only within 2 dB peak signal-to-noise ratio (PSNR) of the performance of a single codec trained with all data sources. We experiment distributed sources with different correlations and show how our data-driven methodology well matches the Slepian- Wolf Theorem in Distributed Source Coding (DSC). To the best of our knowledge, this is the first data-driven DSC framework for general distributed code design with deep learning. 1 Introduction It has been shown by a variety of previous works that deep neural networks (DNN) can achieve comparable results as classical image compression techniques [1–9]. Most of these methods are based on autoencoder networks and quantization of bottleneck representa- tions. These models usually rely on entropy codec to further compress codes. Moreover, to achieve different compression rates it is unavoidable to train multiple models with different regularization parameters separately, which is often computationally intensive.
    [Show full text]
  • The Basic Principles of Data Compression
    The Basic Principles of Data Compression Author: Conrad Chung, 2BrightSparks Introduction Internet users who download or upload files from/to the web, or use email to send or receive attachments will most likely have encountered files in compressed format. In this topic we will cover how compression works, the advantages and disadvantages of compression, as well as types of compression. What is Compression? Compression is the process of encoding data more efficiently to achieve a reduction in file size. One type of compression available is referred to as lossless compression. This means the compressed file will be restored exactly to its original state with no loss of data during the decompression process. This is essential to data compression as the file would be corrupted and unusable should data be lost. Another compression category which will not be covered in this article is “lossy” compression often used in multimedia files for music and images and where data is discarded. Lossless compression algorithms use statistic modeling techniques to reduce repetitive information in a file. Some of the methods may include removal of spacing characters, representing a string of repeated characters with a single character or replacing recurring characters with smaller bit sequences. Advantages/Disadvantages of Compression Compression of files offer many advantages. When compressed, the quantity of bits used to store the information is reduced. Files that are smaller in size will result in shorter transmission times when they are transferred on the Internet. Compressed files also take up less storage space. File compression can zip up several small files into a single file for more convenient email transmission.
    [Show full text]
  • Lzw Compression and Decompression
    LZW COMPRESSION AND DECOMPRESSION December 4, 2015 1 Contents 1 INTRODUCTION 3 2 CONCEPT 3 3 COMPRESSION 3 4 DECOMPRESSION: 4 5 ADVANTAGES OF LZW: 6 6 DISADVANTAGES OF LZW: 6 2 1 INTRODUCTION LZW stands for Lempel-Ziv-Welch. This algorithm was created in 1984 by these people namely Abraham Lempel, Jacob Ziv, and Terry Welch. This algorithm is very simple to implement. In 1977, Lempel and Ziv published a paper on the \sliding-window" compression followed by the \dictionary" based compression which were named LZ77 and LZ78, respectively. later, Welch made a contri- bution to LZ78 algorithm, which was then renamed to be LZW Compression algorithm. 2 CONCEPT Many files in real time, especially text files, have certain set of strings that repeat very often, for example " The ","of","on"etc., . With the spaces, any string takes 5 bytes, or 40 bits to encode. But what if we need to add the whole string to the list of characters after the last one, at 256. Then every time we came across the string like" the ", we could send the code 256 instead of 32,116,104 etc.,. This would take 9 bits instead of 40bits. This is the algorithm of LZW compression. It starts with a "dictionary" of all the single character with indexes from 0 to 255. It then starts to expand the dictionary as information gets sent through. Pretty soon, all the strings will be encoded as a single bit, and compression would have occurred. LZW compression replaces strings of characters with single codes. It does not analyze the input text.
    [Show full text]
  • Claude Elwood Shannon (1916–2001) Solomon W
    Claude Elwood Shannon (1916–2001) Solomon W. Golomb, Elwyn Berlekamp, Thomas M. Cover, Robert G. Gallager, James L. Massey, and Andrew J. Viterbi Solomon W. Golomb Done in complete isolation from the community of population geneticists, this work went unpublished While his incredibly inventive mind enriched until it appeared in 1993 in Shannon’s Collected many fields, Claude Shannon’s enduring fame will Papers [5], by which time its results were known surely rest on his 1948 work “A mathematical independently and genetics had become a very theory of communication” [7] and the ongoing rev- different subject. After his Ph.D. thesis Shannon olution in information technology it engendered. wrote nothing further about genetics, and he Shannon, born April 30, 1916, in Petoskey, Michi- expressed skepticism about attempts to expand gan, obtained bachelor’s degrees in both mathe- the domain of information theory beyond the matics and electrical engineering at the University communications area for which he created it. of Michigan in 1936. He then went to M.I.T., and Starting in 1938 Shannon worked at M.I.T. with after spending the summer of 1937 at Bell Tele- Vannevar Bush’s “differential analyzer”, the an- phone Laboratories, he wrote one of the greatest cestral analog computer. After another summer master’s theses ever, published in 1938 as “A sym- (1940) at Bell Labs, he spent the academic year bolic analysis of relay and switching circuits” [8], 1940–41 working under the famous mathemati- in which he showed that the symbolic logic of cian Hermann Weyl at the Institute for Advanced George Boole’s nineteenth century Laws of Thought Study in Princeton, where he also began thinking provided the perfect mathematical model for about recasting communications on a proper switching theory (and indeed for the subsequent mathematical foundation.
    [Show full text]
  • Digital Communication Systems 2.2 Optimal Source Coding
    Digital Communication Systems EES 452 Asst. Prof. Dr. Prapun Suksompong [email protected] 2. Source Coding 2.2 Optimal Source Coding: Huffman Coding: Origin, Recipe, MATLAB Implementation 1 Examples of Prefix Codes Nonsingular Fixed-Length Code Shannon–Fano code Huffman Code 2 Prof. Robert Fano (1917-2016) Shannon Award (1976 ) Shannon–Fano Code Proposed in Shannon’s “A Mathematical Theory of Communication” in 1948 The method was attributed to Fano, who later published it as a technical report. Fano, R.M. (1949). “The transmission of information”. Technical Report No. 65. Cambridge (Mass.), USA: Research Laboratory of Electronics at MIT. Should not be confused with Shannon coding, the coding method used to prove Shannon's noiseless coding theorem, or with Shannon–Fano–Elias coding (also known as Elias coding), the precursor to arithmetic coding. 3 Claude E. Shannon Award Claude E. Shannon (1972) Elwyn R. Berlekamp (1993) Sergio Verdu (2007) David S. Slepian (1974) Aaron D. Wyner (1994) Robert M. Gray (2008) Robert M. Fano (1976) G. David Forney, Jr. (1995) Jorma Rissanen (2009) Peter Elias (1977) Imre Csiszár (1996) Te Sun Han (2010) Mark S. Pinsker (1978) Jacob Ziv (1997) Shlomo Shamai (Shitz) (2011) Jacob Wolfowitz (1979) Neil J. A. Sloane (1998) Abbas El Gamal (2012) W. Wesley Peterson (1981) Tadao Kasami (1999) Katalin Marton (2013) Irving S. Reed (1982) Thomas Kailath (2000) János Körner (2014) Robert G. Gallager (1983) Jack KeilWolf (2001) Arthur Robert Calderbank (2015) Solomon W. Golomb (1985) Toby Berger (2002) Alexander S. Holevo (2016) William L. Root (1986) Lloyd R. Welch (2003) David Tse (2017) James L.
    [Show full text]
  • Principles of Communications ECS 332
    Principles of Communications ECS 332 Asst. Prof. Dr. Prapun Suksompong (ผศ.ดร.ประพันธ ์ สขสมปองุ ) [email protected] 1. Intro to Communication Systems Office Hours: Check Google Calendar on the course website. Dr.Prapun’s Office: 6th floor of Sirindhralai building, 1 BKD 2 Remark 1 If the downloaded file crashed your device/browser, try another one posted on the course website: 3 Remark 2 There is also three more sections from the Appendices of the lecture notes: 4 Shannon's insight 5 “The fundamental problem of communication is that of reproducing at one point either exactly or approximately a message selected at another point.” Shannon, Claude. A Mathematical Theory Of Communication. (1948) 6 Shannon: Father of the Info. Age Documentary Co-produced by the Jacobs School, UCSD- TV, and the California Institute for Telecommunic ations and Information Technology 7 [http://www.uctv.tv/shows/Claude-Shannon-Father-of-the-Information-Age-6090] [http://www.youtube.com/watch?v=z2Whj_nL-x8] C. E. Shannon (1916-2001) Hello. I'm Claude Shannon a mathematician here at the Bell Telephone laboratories He didn't create the compact disc, the fax machine, digital wireless telephones Or mp3 files, but in 1948 Claude Shannon paved the way for all of them with the Basic theory underlying digital communications and storage he called it 8 information theory. C. E. Shannon (1916-2001) 9 https://www.youtube.com/watch?v=47ag2sXRDeU C. E. Shannon (1916-2001) One of the most influential minds of the 20th century yet when he died on February 24, 2001, Shannon was virtually unknown to the public at large 10 C.
    [Show full text]
  • Marconi Society - Wikipedia
    9/23/2019 Marconi Society - Wikipedia Marconi Society The Guglielmo Marconi International Fellowship Foundation, briefly called Marconi Foundation and currently known as The Marconi Society, was established by Gioia Marconi Braga in 1974[1] to commemorate the centennial of the birth (April 24, 1874) of her father Guglielmo Marconi. The Marconi International Fellowship Council was established to honor significant contributions in science and technology, awarding the Marconi Prize and an annual $100,000 grant to a living scientist who has made advances in communication technology that benefits mankind. The Marconi Fellows are Sir Eric A. Ash (1984), Paul Baran (1991), Sir Tim Berners-Lee (2002), Claude Berrou (2005), Sergey Brin (2004), Francesco Carassa (1983), Vinton G. Cerf (1998), Andrew Chraplyvy (2009), Colin Cherry (1978), John Cioffi (2006), Arthur C. Clarke (1982), Martin Cooper (2013), Whitfield Diffie (2000), Federico Faggin (1988), James Flanagan (1992), David Forney, Jr. (1997), Robert G. Gallager (2003), Robert N. Hall (1989), Izuo Hayashi (1993), Martin Hellman (2000), Hiroshi Inose (1976), Irwin M. Jacobs (2011), Robert E. Kahn (1994) Sir Charles Kao (1985), James R. Killian (1975), Leonard Kleinrock (1986), Herwig Kogelnik (2001), Robert W. Lucky (1987), James L. Massey (1999), Robert Metcalfe (2003), Lawrence Page (2004), Yash Pal (1980), Seymour Papert (1981), Arogyaswami Paulraj (2014), David N. Payne (2008), John R. Pierce (1979), Ronald L. Rivest (2007), Arthur L. Schawlow (1977), Allan Snyder (2001), Robert Tkach (2009), Gottfried Ungerboeck (1996), Andrew Viterbi (1990), Jack Keil Wolf (2011), Jacob Ziv (1995). In 2015, the prize went to Peter T. Kirstein for bringing the internet to Europe. Since 2008, Marconi has also issued the Paul Baran Marconi Society Young Scholar Awards.
    [Show full text]
  • MÁSTER UNIVERSITARIO EN INGENIERÍA DE TELECOMUNICACIÓN TRABAJO FIN DE MASTER Analysis and Implementation of New Techniques Fo
    MÁSTER UNIVERSITARIO EN INGENIERÍA DE TELECOMUNICACIÓN TRABAJO FIN DE MASTER Analysis and Implementation of New Techniques for the Enhancement of the Reliability and Delay of Haptic Communications LUIS AMADOR GONZÁLEZ 2017 MASTER UNIVERSITARIO EN INGENIERÍA DE TELECOMUNICACIÓN TRABAJO FIN DE MASTER Título: Analysis and Implementation of New Techniques for the Enhancement of the Reliability and Delay of Haptic Communications Autor: D. Luis Amador González Tutor: D. Hamid Aghvami Ponente: D. Santiago Zazo Bello Departamento: MIEMBROS DEL TRIBUNAL Presidente: D. Vocal: D. Secretario: D. Suplente: D. Los miembros del tribunal arriba nombrados acuerdan otorgar la calificación de: ……… Madrid, a de de 20… UNIVERSIDAD POLITÉCNICA DE MADRID ESCUELA TÉCNICA SUPERIOR DE INGENIEROS DE TELECOMUNICACIÓN MASTER UNIVERSITARIO EN INGENIERÍA DE TELECOMUNICACIÓN TRABAJO FIN DE MASTER Analysis and Implementation of New Techniques for the Enhancement of the Reliability and Delay of Haptic Communications LUIS AMADOR GONZÁLEZ 2017 CONTENTS LIST OF FIGURES .....................................................................................2 LIST OF TABLES ......................................................................................3 ABBREVIATIONS .....................................................................................4 SUMMARY ............................................................................................5 RESUMEN .............................................................................................6 ACKNOWLEDGEMENTS
    [Show full text]
  • Channel Coding
    1 Channel Coding: The Road to Channel Capacity Daniel J. Costello, Jr., Fellow, IEEE, and G. David Forney, Jr., Fellow, IEEE Submitted to the Proceedings of the IEEE First revision, November 2006 Abstract Starting from Shannon’s celebrated 1948 channel coding theorem, we trace the evolution of channel coding from Hamming codes to capacity-approaching codes. We focus on the contributions that have led to the most significant improvements in performance vs. complexity for practical applications, particularly on the additive white Gaussian noise (AWGN) channel. We discuss algebraic block codes, and why they did not prove to be the way to get to the Shannon limit. We trace the antecedents of today’s capacity-approaching codes: convolutional codes, concatenated codes, and other probabilistic coding schemes. Finally, we sketch some of the practical applications of these codes. Index Terms Channel coding, algebraic block codes, convolutional codes, concatenated codes, turbo codes, low-density parity- check codes, codes on graphs. I. INTRODUCTION The field of channel coding started with Claude Shannon’s 1948 landmark paper [1]. For the next half century, its central objective was to find practical coding schemes that could approach channel capacity (hereafter called “the Shannon limit”) on well-understood channels such as the additive white Gaussian noise (AWGN) channel. This goal proved to be challenging, but not impossible. In the past decade, with the advent of turbo codes and the rebirth of low-density parity-check codes, it has finally been achieved, at least in many cases of practical interest. As Bob McEliece observed in his 2004 Shannon Lecture [2], the extraordinary efforts that were required to achieve this objective may not be fully appreciated by future historians.
    [Show full text]