Error Control Techniques for Interactive Low-Bit Rate Video Transmission

Total Page:16

File Type:pdf, Size:1020Kb

Error Control Techniques for Interactive Low-Bit Rate Video Transmission Error Control Techniques for Interactive Low-bit Rate Video Transmission over the Internet Injong Rhee Department of Computer Science North Carolina State University Raleigh, NC 27695-7534 E-mail: [email protected] Abstract The e ect of packet loss is also far-reaching. Packet loss degrades not only the quality of the frames con- tained in lost packets, but also the quality of their sub- A new retransmission-based error control technique is sequent frames. This is b ecause of motion comp ensation presented that does not incur any additional latency in employed by video co decs. Motion-comp ensated video frame playout times, and hence are suitable for inter- co decs remove temp oral redundancy in a video stream active applications. It takes advantage of the motion by enco ding only pixel value di erence prediction er- prediction loop employed in most motion compensation- ror b etween an image to b e enco ded inter-frame and a basedcodecs. By correcting errors in a referenceframe previously transmitted image reference frame. Packet causedbyearlier packet loss, it prevents error propaga- loss can intro duce an error in a reference frame, which tion. The technique rearranges the temporal dependency can b e propagated to its subsequent frames and get am- of frames so that a displayedframe is referenced for the pli ed as more packets are lost. decoding of its succeeding dependent frames much later than its display time. Thus, the delay in repairing lost Error propagation can be controlled by more fre- packets can be e ectively masked out. The developed quently adding intra frames which are co ded temp o- technique is combined with layered videocoding to main- rally indep endently. However, the ratio of the com- tain consistently good video quality even under heavy pression eciency of an intra-frame over an inter-frame packet loss. Through the results of extensive Internet is as large as 3 to 6 times. Increasing the frequency experiments, the paper shows that layeredcoding can be of intra-frames could increase the bandwidth require- very e ective when combined with the retransmission- menttoomuch for video transmission over a bandwidth- based error control technique for low-bit rate transmis- constraint network. Nonetheless, the severe degradation sion over best e ort networks where no network-level of image quality due to error propagation has forced mechanism exists for protecting high priority data from several p opular video conferencing to ols, suchas nv[7 ], packet loss. vic[13 ] and CU-SeeMe[6 ], to adopt an even more drastic approach. Using a technique called conditional replen- ishment, these to ols lter out the blo cks that have not 1 Intro duction changed much from the previous frame and intra-co de the remaining blo cks. Since all the co ded blo cks are temp orally indep endent, packet loss a ects only those Transmitting interactive video over packet switching frames contained in lost packets. However, this en- networks is time-constrained. If a packet containing a hanced error resilience comes at the cost of low com- video frame is not delivered b efore its scheduled play- pression eciency. Additional compression can always out time, and the frame cannot be displayed and ac- be obtained if temp oral redundancy is removed from cordingly, the overall quality of video degrades. The each co ded blo ck i.e., by co ding only their prediction diculty of meeting this time constraint lies mainly in error. frequent o ccurrences of packet loss which are commonly caused by network congestion. The recovery of lost The goal is to develop an error recovery scheme packets b efore their display times is not always feasi- that solves the error propagation problem without los- ble b ecause of the delay asso ciated with detecting and ing much compression eciency. Retransmission-based repairing the lost packets. and forward error correction schemes are the two ma jor error control schemes found in the literature. Retransmission-based error recovery REC can pro- vide go o d error resilience without incurring much band- width overhead b ecause packets are retransmitted only when some indications exist that they are lost. However, retransmission always involves additional transmission delay, and thus has b een widely known ine ective for in- teractive real-time video applications such as video con- frames temp orally dep end only on the essential signals ferencing. Many researchers prop osed use of extended of their reference frames. This dep endency ensures that control or playout times to allow retransmitted packets the distortion of the enhancement signals of a frame to arrive in time for display [21 , 18 , 12 , 23 ]. This implies do es not a ect the quality of its succeeding frames. that the playout time of a frame is delayed by at least Since essential signals are protected by FEC, this dep en- three one-way trip times after its initial transmission dency will signi cantly reduce error propagation. How- two for packet transmissions and one for a retransmis- ever, QAL has its own limitations. First, since frames sion request. Under the currentInternet environment, are motion-comp ensated only to the essential signals of this delaywould b e intolerable for interactive video ap- their reference frames, the temp oral redundancy present plications, and can severely reduce the interactivityof in the enhancement signals are not exploited at all, re- the application. sulting in low compression eciency. Second, under heavy packet loss, even essential signals can be lost, Retransmission, however, can be still a very e ec- causing error propagation. tive technique to improve error resilience in interactive real-time video conferencing. In this pap er, we study a These limitations, however, can b e greatly reduced REC scheme, called periodic temporal dependency dis- when QAL is combined with PTDD. In this pap er, we tance PTDD, that can be used for interactive video show that a motion-comp ensated layered co ding tech- applications. In particular, the schemes do not require nique such as QAL can b e very e ective for transmis- any arti cial extension of control time and play-out de- sion over a lossy packet-switched network even when no lays, and thus are suitable for interactive applications. network-level mechanism to protect a prioritized stream In the scheme, frames are simply displayed at their nor- exists. When used with PTDD, the layering technique mal playout times without any delay, as they are de- yields reasonably go o d compression eciency since p e- co ded. Thus, if a packet arrives after the playout time rio dic frames temp orally dep ends on b oth the essential of its frame, the frame will be displayed with errors. and enhancement signals of their reference frames. This However, this \late" packet can b e used to remove error dep endency is safe b ecause retransmission can recover propagation. Rather than discarding the late packet, lost packets containing b oth signals. In addition, p eri- PTDD uses it to restore its frame although the frame o dic frames reduce error propagation b ecause their im- is displayed. Because the frame is used as a reference mediately succeeding non-p erio dic frames use them as frame for its succeeding frames, restoring the reference a reference frame. frame stops error propagation. We conducted b oth transatlantic and transpaci c In- To allow enough time for retransmitted packets to ternet transmission tests from the East Coast of the arrive b efore their frames are referenced for the recon- US to show the utility of the REC-based techniques for struction of their dep endent frames, the PTDD scheme transmission over a long distance network. The exp er- extends the temporal dependency distance TDD of imental results indicated that the prop osed technique frames. The TDD of a frame is de ned to b e the min- achieve go o d error resilience and compression eciency. imum numb er of frame intervals or inter-frame delay In Section 2, we describ e the related work, and in between that frame and another frame on which it is Section 3 we describ e PTDD and QAL schemes. Section temp orally dep endent. In the PTDD scheme, every p- 4 contains a discussion of the exp erimental results, and th frame whichwe call periodic frame has an extended Section 5 contains the conclusion. TDD while the other frames have TDD 1. Note that the extension of TDD do es not a ect the playout times of frames since the playout times of frames are deter- 2 Related Work mined solely byinter-frame delays and the initial con- trol time which is used to reduce delay jitter. In the PTDD scheme, the TDD of p erio dic frames is deter- We rst discuss the related work on retransmission- mined by the estimated delaybetween the sender and based techniques for video transmission over packet- the receiver. switched networks and then discuss related work on lay- ered video co ding. In PTDD, non-p erio dic frames with TDD 1 are not protected by retransmission. If packet loss o ccurs for a Retransmission-based error recovery. Dempsey et al.[5 ] non-p erio dic frame, the frame will b e displayed with er- applied retransmission for the recovery of audio packets. rors which can b e propagated to its subsequent frames They showed that by adding some delay b efore the play- until the next p erio dic frame is received. The receiv- out of each received audio packet, retransmission can ing video quality will p erio dically uctuate. To remedy b e used to protect audio data from packet loss.
Recommended publications
  • Video Codec Requirements and Evaluation Methodology
    Video Codec Requirements 47pt 30pt and Evaluation Methodology Color::white : LT Medium Font to be used by customers and : Arial www.huawei.com draft-filippov-netvc-requirements-01 Alexey Filippov, Huawei Technologies 35pt Contents Font to be used by customers and partners : • An overview of applications • Requirements 18pt • Evaluation methodology Font to be used by customers • Conclusions and partners : Slide 2 Page 2 35pt Applications Font to be used by customers and partners : • Internet Protocol Television (IPTV) • Video conferencing 18pt • Video sharing Font to be used by customers • Screencasting and partners : • Game streaming • Video monitoring / surveillance Slide 3 35pt Internet Protocol Television (IPTV) Font to be used by customers and partners : • Basic requirements: . Random access to pictures 18pt Random Access Period (RAP) should be kept small enough (approximately, 1-15 seconds); Font to be used by customers . Temporal (frame-rate) scalability; and partners : . Error robustness • Optional requirements: . resolution and quality (SNR) scalability Slide 4 35pt Internet Protocol Television (IPTV) Font to be used by customers and partners : Resolution Frame-rate, fps Picture access mode 2160p (4K),3840x2160 60 RA 18pt 1080p, 1920x1080 24, 50, 60 RA 1080i, 1920x1080 30 (60 fields per second) RA Font to be used by customers and partners : 720p, 1280x720 50, 60 RA 576p (EDTV), 720x576 25, 50 RA 576i (SDTV), 720x576 25, 30 RA 480p (EDTV), 720x480 50, 60 RA 480i (SDTV), 720x480 25, 30 RA Slide 5 35pt Video conferencing Font to be used by customers and partners : • Basic requirements: . Delay should be kept as low as possible 18pt The preferable and maximum delay values should be less than 100 ms and 350 ms, respectively Font to be used by customers .
    [Show full text]
  • (A/V Codecs) REDCODE RAW (.R3D) ARRIRAW
    What is a Codec? Codec is a portmanteau of either "Compressor-Decompressor" or "Coder-Decoder," which describes a device or program capable of performing transformations on a data stream or signal. Codecs encode a stream or signal for transmission, storage or encryption and decode it for viewing or editing. Codecs are often used in videoconferencing and streaming media solutions. A video codec converts analog video signals from a video camera into digital signals for transmission. It then converts the digital signals back to analog for display. An audio codec converts analog audio signals from a microphone into digital signals for transmission. It then converts the digital signals back to analog for playing. The raw encoded form of audio and video data is often called essence, to distinguish it from the metadata information that together make up the information content of the stream and any "wrapper" data that is then added to aid access to or improve the robustness of the stream. Most codecs are lossy, in order to get a reasonably small file size. There are lossless codecs as well, but for most purposes the almost imperceptible increase in quality is not worth the considerable increase in data size. The main exception is if the data will undergo more processing in the future, in which case the repeated lossy encoding would damage the eventual quality too much. Many multimedia data streams need to contain both audio and video data, and often some form of metadata that permits synchronization of the audio and video. Each of these three streams may be handled by different programs, processes, or hardware; but for the multimedia data stream to be useful in stored or transmitted form, they must be encapsulated together in a container format.
    [Show full text]
  • Netflix and the Development of the Internet Television Network
    Syracuse University SURFACE Dissertations - ALL SURFACE May 2016 Netflix and the Development of the Internet Television Network Laura Osur Syracuse University Follow this and additional works at: https://surface.syr.edu/etd Part of the Social and Behavioral Sciences Commons Recommended Citation Osur, Laura, "Netflix and the Development of the Internet Television Network" (2016). Dissertations - ALL. 448. https://surface.syr.edu/etd/448 This Dissertation is brought to you for free and open access by the SURFACE at SURFACE. It has been accepted for inclusion in Dissertations - ALL by an authorized administrator of SURFACE. For more information, please contact [email protected]. Abstract When Netflix launched in April 1998, Internet video was in its infancy. Eighteen years later, Netflix has developed into the first truly global Internet TV network. Many books have been written about the five broadcast networks – NBC, CBS, ABC, Fox, and the CW – and many about the major cable networks – HBO, CNN, MTV, Nickelodeon, just to name a few – and this is the fitting time to undertake a detailed analysis of how Netflix, as the preeminent Internet TV networks, has come to be. This book, then, combines historical, industrial, and textual analysis to investigate, contextualize, and historicize Netflix's development as an Internet TV network. The book is split into four chapters. The first explores the ways in which Netflix's development during its early years a DVD-by-mail company – 1998-2007, a period I am calling "Netflix as Rental Company" – lay the foundations for the company's future iterations and successes. During this period, Netflix adapted DVD distribution to the Internet, revolutionizing the way viewers receive, watch, and choose content, and built a brand reputation on consumer-centric innovation.
    [Show full text]
  • The H.264 Advanced Video Coding (AVC) Standard
    Whitepaper: The H.264 Advanced Video Coding (AVC) Standard What It Means to Web Camera Performance Introduction A new generation of webcams is hitting the market that makes video conferencing a more lifelike experience for users, thanks to adoption of the breakthrough H.264 standard. This white paper explains some of the key benefits of H.264 encoding and why cameras with this technology should be on the shopping list of every business. The Need for Compression Today, Internet connection rates average in the range of a few megabits per second. While VGA video requires 147 megabits per second (Mbps) of data, full high definition (HD) 1080p video requires almost one gigabit per second of data, as illustrated in Table 1. Table 1. Display Resolution Format Comparison Format Horizontal Pixels Vertical Lines Pixels Megabits per second (Mbps) QVGA 320 240 76,800 37 VGA 640 480 307,200 147 720p 1280 720 921,600 442 1080p 1920 1080 2,073,600 995 Video Compression Techniques Digital video streams, especially at high definition (HD) resolution, represent huge amounts of data. In order to achieve real-time HD resolution over typical Internet connection bandwidths, video compression is required. The amount of compression required to transmit 1080p video over a three megabits per second link is 332:1! Video compression techniques use mathematical algorithms to reduce the amount of data needed to transmit or store video. Lossless Compression Lossless compression changes how data is stored without resulting in any loss of information. Zip files are losslessly compressed so that when they are unzipped, the original files are recovered.
    [Show full text]
  • A Comparison of Video Formats for Online Teaching Ross A
    Contemporary Issues in Education Research – First Quarter 2017 Volume 10, Number 1 A Comparison Of Video Formats For Online Teaching Ross A. Malaga, Montclair State University, USA Nicole B. Koppel, Montclair State University, USA ABSTRACT The use of video to deliver content to students online has become increasingly popular. However, educators are often plagued with the question of which format to use to deliver asynchronous video material. Whether it is a College or University committing to a common video format or an individual instructor selecting the method that works best for his or her course, this research presents a comparison of various video formats that can be applied to online education and provides guidance in which one to select. Keywords: Online Teaching; Video Formats; Technology Acceptance Model INTRODUCTION istance learning is one of the most talked-about topics in higher education today. Online and hybrid (or blended) learning removes location and time-bound constraints of the traditional college classroom to a learning environment that can occur anytime or anywhere in a global environment. DAccording to research by the Online Learning Consortium, over 5 million students took an online course in the Fall 2014 semester. This represents an increase in online enrollment of over 3.9% in just one year. In 2014, 28% of higher education students took one or more courses online (Allen, I. E. and Seaman, J, 2016). With this incredible growth, albeit slower than the growth in previous years, institutions of higher education are continuing to increase their online course and program offerings. As such, institutions need to find easy to develop, easy to use, reliable, and reasonably priced technologies to deliver online content.
    [Show full text]
  • WILL Youtube SAIL INTO the DMCA's SAFE HARBOR OR SINK for INTERNET PIRACY?
    THE JOHN MARSHALL REVIEW OF INTELLECTUAL PROPERTY LAW WILL YoUTUBE SAIL INTO THE DMCA's SAFE HARBOR OR SINK FOR INTERNET PIRACY? MICHAEL DRISCOLL ABSTRACT Is YouTube, the popular video sharing website, a new revolution in information sharing or a profitable clearing-house for unauthorized distribution of copyrighted material? YouTube's critics claim that it falls within the latter category, in line with Napster and Grokster. This comment, however, determines that YouTube is fundamentally different from past infringers in that it complies with statutory provisions concerning the removal of copyrighted materials. Furthermore, YouTube's central server architecture distinguishes it from peer-to-peer file sharing websites. This comment concludes that any comparison to Napster or Grokster is superficial, and overlooks the potential benefits of YouTube to copyright owners and to society at large. Copyright © 2007 The John Marshall Law School Cite as Michael Driscoll, Will YouTube Sail into the DMCA's Safe Harboror Sink for Internet Piracy?, 6 J. MARSHALL REV. INTELL. PROP. L. 550 (2007). WILL YoUTUBE SAIL INTO THE DMCA's SAFE HARBOR OR SINK FOR INTERNET PIRACY? MICHAEL DRISCOLL* 'A sorry agreement is better than a good suit in law." English Proverb1 INTRODUCTION The year 2006 proved a banner year for YouTube, Inc. ("YouTube"), a well- known Internet video sharing service, so much so that Time Magazine credited YouTube for making "You" the Person of the Year. 2 Despite this seemingly positive development for such a young company, the possibility of massive copyright 3 infringement litigation looms over YouTube's future. For months, YouTube was walking a virtual tightrope by obtaining licensing agreements with major copyright owners, yet increasingly gaining popularity through its endless video selection, both legal and otherwise.
    [Show full text]
  • MPEG Standardization Roadmap October 2017 Version Why a Standardisation Roadmap?
    MPEG Standardization Roadmap October 2017 Version Why a Standardisation Roadmap? • MPEG has created, and is still producing, media standards that enable huge markets to flourish • MPEG works on requirements from industry • Many industries represented in MPEG, but not all of MPEG’s customers can or need to participate in the process • MPEG wants to inform its customers about its long- term plans (~ 5 years out) • … and collect feedback and requirements from these customers • … including in this session What is in the Roadmap • Our roadmap is a short document. • It briefly outlines MPEG’s most important standards MPEG Standards 1990 1995 2000 2005 2010 2015 Digital Music Custom Fonts on Television Distribution Web & DigiPub Internet Mobile Video HD Distribution UHD & Immersive Audio Services MPEG-2 MP3 Video & MPEG-4 AAC AVC OFF HEVC 3D Audio Systems Video MP4 FF DASH MMT MLAF Media Storage and Unified Media Second Streaming Streaming Screen New Forms of Digital Television What is in the Roadmap • Our roadmap is a short document. • It briefly outlines MPEG’s most important standards • It then gives an overview of MPEG’s activities Multimedia Platform MPEG’s Areas Systems, video, Technologies MPEG- audio and U,V of Activity transport MPEG- E,M Technologies MPEG- for content B,C,D, Device and e-commerce application DASH interfaces MPEG-A Multimedia Application Compression of Formats (combinations video, audio MPEG-21 of content formats) and 3DG MPEG-7 Description of video, MPEG- audio and multimedia 1,2,4,H,I for content search What is in the Roadmap • Our roadmap is a short document.
    [Show full text]
  • Reliable Streaming Media Delivery
    WHITE PAPER Reliable Streaming Media Delivery www.ixiacom.com 915-1777-01 Rev B January 2014 2 Table of Contents Overview ................................................................................................................ 4 Business Trends ..................................................................................................... 6 Technology Trends ................................................................................................. 7 Building in Reliability .............................................................................................. 8 Ixia’s Test Solutions ............................................................................................... 9 Streaming Media Test Scenarios ........................................................................... 11 Conclusion .............................................................................................................13 3 Overview Streaming media is an evolving set of technologies that deliver multimedia content over the Internet and private networks. A number of service businesses are dedicated to streaming media delivery, including YouTube, Brightcove, Vimeo, Metacafe, BBC iPlayer, and Hulu. Streaming video delivery is growing dramatically: according to the comScore Video Metrix1, Americans viewed a significantly higher number of videos in 2009 than in 2008 (up 19%) due to both increased content consumption and the growing number of video ads delivered. In January of 2010, more than 170 million viewers watched videos
    [Show full text]
  • A Study on H.26X Family of Video Streaming Compression Techniques
    International Journal of Pure and Applied Mathematics Volume 117 No. 10 2017, 63-66 ISSN: 1311-8080 (printed version); ISSN: 1314-3395 (on-line version) url: http://www.ijpam.eu doi: 10.12732/ijpam.v117i10.12 Special Issue ijpam.eu A Study On H.26x Family Of Video Streaming Compression Techniques Farooq Sunar Mahammad, V Madhu Viswanatham School of Computing Science & Engineering VIT University Vellore, Tamil Nadu, India [email protected] [email protected] Abstract— H.264 is a standard was one made in collaboration MPEG-4 Part 10 (AVC), MPEG21,MPEG-7 and M-JPEG. with ITU-T and ISO/IEC Moving Picture Expert Group. The ITU-I VCEG standard incorporates H.26x series, H.264, standard is turning out to be more prevalent as the principle H.261 and H.263. In this paper, distinctive video weight objective of H.264 is to accomplish spatial versatility and to strategies are kept an eye on, starting from H.261 series to the enhance the compression execution contrasted with different most recent such standard, known as H.265/HEVC. Fig. 1 standards.The H.264 is utilized as a part of spatial way to encode the video, so that the size is reduced and the quantity of the expounds the evolution of MPEG standards and ITU-T frames is being decreased and it’s in this way it accomplish Recommendations [2]. versatility. It gives new degree to making higher quality of video encoders and also decoders that give extensive level of quality video streams at kept up bit-rates (contrasted with past standdards), or, on the other hand, a similar quality video at a lower bitrate.
    [Show full text]
  • Internet Video Telephony Allows Speech Reading by Deaf Individuals and Improves Speech Perception by Cochlear Implant Users
    Internet Video Telephony Allows Speech Reading by Deaf Individuals and Improves Speech Perception by Cochlear Implant Users Georgios Mantokoudis*, Claudia Da¨hler, Patrick Dubach, Martin Kompis, Marco D. Caversaccio, Pascal Senn University Department of Otorhinolaryngology, Head & Neck Surgery, Inselspital, Bern, Switzerland Abstract Objective: To analyze speech reading through Internet video calls by profoundly hearing-impaired individuals and cochlear implant (CI) users. Methods: Speech reading skills of 14 deaf adults and 21 CI users were assessed using the Hochmair Schulz Moser (HSM) sentence test. We presented video simulations using different video resolutions (12806720, 6406480, 3206240, 1606120 px), frame rates (30, 20, 10, 7, 5 frames per second (fps)), speech velocities (three different speakers), webcameras (Logitech Pro9000, C600 and C500) and image/sound delays (0–500 ms). All video simulations were presented with and without sound and in two screen sizes. Additionally, scores for live SkypeTM video connection and live face-to-face communication were assessed. Results: Higher frame rate (.7 fps), higher camera resolution (.6406480 px) and shorter picture/sound delay (,100 ms) were associated with increased speech perception scores. Scores were strongly dependent on the speaker but were not influenced by physical properties of the camera optics or the full screen mode. There is a significant median gain of +8.5%pts (p = 0.009) in speech perception for all 21 CI-users if visual cues are additionally shown. CI users with poor open set speech perception scores (n = 11) showed the greatest benefit under combined audio-visual presentation (median speech perception +11.8%pts, p = 0.032). Conclusion: Webcameras have the potential to improve telecommunication of hearing-impaired individuals.
    [Show full text]
  • The Impact of Online Video Distribution on the Global Market for Digital Content
    The Impact of Online Video Distribution on the Global Market for Digital Content David Blackburn, Ph.D. Jeffrey A. Eisenach, Ph.D. Bruno Soria, Ph.D. March 2019 About the Authors Dr. Blackburn is a Director in NERA’s Communications, Media, and Internet Practice as well its Intellectual Property and Antitrust Practices. Among other issues, Dr. Blackburn’s work at NERA has focused on media production and distribution, and assessing the value of IP in music, television, and film. Dr. Blackburn has taught at the undergraduate level at Harvard University and Framingham State College, and at the graduate level at the Universidad Nacional de Tucumán in Argentina. Dr. Eisenach is a Managing Director and Co-Chair of NERA’s Communications, Media, and Internet Practice. He is also an Adjunct Professor at George Mason University Law School, where he teaches Regulated Industries, and a Visiting Scholar at the American Enterprise Institute, where he focuses on policies affecting the information technology sector, innovation, and entrepreneurship. Previously, Dr. Eisenach served in senior policy positions at the U.S. Federal Trade Commission and the White House Office of Management and Budget, and on the faculties of Harvard University’s Kennedy School of Government and Virginia Polytechnic Institute and State University. Dr. Soria is an Associate Director in NERA’s Communications, Media and Internet Practice. While at NERA, he has advised governments, telecommunications operators and media companies, including on convergent competition and the pricing of content. He is also Guest Professor at the University of Barcelona where he lectures on Telecommunications Economics and Regulation. Previously, he held executive positions in Telefónica and MCI Worldcom.
    [Show full text]
  • History Analog Video Transmission
    Videotelephony - AccessScience from McGraw-Hill Education http://www.accessscience.com/content/videotelephony/732100 (http://www.accessscience.com/) Article by: Bleiweis, John J. JJB Associates, Great Falls, Virginia. Publication year: 2014 DOI: http://dx.doi.org/10.1036/1097-8542.732100 (http://dx.doi.org/10.1036/1097-8542.732100) Content History Infrastructure Bibliography Analog video transmission Video telephone hardware Additional Readings Digital video transmission Video telephone software A means of simultaneous, two-way communication comprising both audio and video elements. Participants in a video telephone call can both see and hear each other in real time. Videotelephony is a subset of teleconferencing, broadly defined as the various ways and means by which people communicate with one another over some distance. Initially conceived as an extension to the telephone, videotelephony is now possible using computers with network connections. In addition to general personal use, there are specific professional applications, such as criminal justice, health care delivery, and surveillance that can greatly benefit from videotelephony. See also: Teleconferencing (/content/teleconferencing/680075) History Although basic research on the technology of videotelephony dates back to 1925, the first public demonstration of the concept was by American Telephone and Telegraph Corporation (AT&T) at the 1964 New York World's Fair. The device was called Picturephone, and the high cost of the analog circuits required to support it made it very expensive and thus unsuitable for the market place. The 1970s brought the first attempts at digitization of transmissions. The video telephones comprised four parts: a standard touch-tone telephone, a small screen with a camera and loudspeaker, a control pad with user controls and a microphone, and a service unit (coder-decoder, or CODEC) that converted the analog signals to digital signals for transmission and vice versa for reception and display.
    [Show full text]