Displayport 1.2 Spec Pdf

Total Page:16

File Type:pdf, Size:1020Kb

Displayport 1.2 Spec Pdf Displayport 1.2 spec pdf Continue Digital Display Interface DisplayPort Type Digital Audio/Video Connector Production HistoryDesigner VESADesigned May 2006Manufacturer VariousProduced 2008-presentSupersed DVI, VGA, SCART, RGB ComponentSupersed by NoneGeneral specsLength Various connected YesExternal YesAudio Signal Optional; 1-8 channels, 16 or 24-bit linear PCM; Sampling speed of 32-192 kHz; Maximum bitrate 36,864 kbps (4608 kB/s) Video signal Optional, maximum resolution is limited to available bandwidth Of 20 pins for external connectors on desktop computers, laptops, graphics cards, monitors, etc. and 30/20 pins for internal connections between graph engines and flat panels. Electric Senal No.3.3 VMax. 16.0 VMax. Current signal 0.5 ADataData YesBitrate 1.62, 2.7, 5.4, 8.1, or 20 Gbit/s bandwidth rate; 1, 2 or 4 lanes; (effective total of 5.184, 8.64, 17.28, 25.92, or 77.37 Gbit/s for a four-lane connection); 2 or 720 Mbps (effectively 1 or 576 Mbps) for the auxiliary channel. Micro-packetPin out External connector (source-side) on PCBPin 1 ML_Lane 0 (p) Lane 0 (positive) Pin 2 GND GroundPin 3 ML_Lane 0 (n) Lane 0 (negative) Pin 4 ML_Lane 1 (p) Lane 1 (p) positive) Pin 5 GND GroundPin 6 ML_Lane 1 (n) Lane 1 (negative) Pin 7 ML_Lane 2 (p) Lane 2 (positive) Pin 8 GND GroundPin 9 ML_Lane 2 (n) Lane 2 (negative) Pin 10 ML_Lane ML_Lane 3 (3 (positive) (p) Lane 3 (positive)Pin 11 GND GroundPin 12 ML_Lane 3 (n) Lane 3 (negative) Pin 13 CONFIG1 Land-Connected b'Pin 14 CONFIG2 Connected to Land b'Pin 15 AUX CH (p) Auxiliary Channel (positive)Pin 16 GND GroundPin 17 AUX CH (n) Auxiliary Channel (negative)Pin 18 Hot Fork Hot Fork detectPin 19 Return for powerPin 20 DP_PWR Power for connector (3.3 V 500 mA) g h This pinout for the side connector source , the pinout of the sink-side connector will have 0'3 lanes reverse in order; i.e. band 3 will be on contact 1 (n) and 3(p), while band 0 will be on contact 10(n) and 12 (p). B Pins 13 and 14 can be either directly connected to the ground or connected to the ground using a retractable device. The DisplayPort A Mini DisplayPort receptacle (centre) with Thunderbolt 3 port (left) and power input (right) DisplayPort (DP) is a digital display interface developed by a consortium of PC and chip manufacturers and standardized by the Video Electronics Standards Association (VESA). The interface is mainly used to connect the video source to a display device, such as a computer monitor, and it can also carry audio, USB and other forms of data. DisplayPort was designed to replace VGA, FPD-Link and digital visual interface (DVI). The back interface is compatible with others such as HDMI and DVI, using active or passive adapters. DisplayPort review is the first display interface to rely on packaged data, a form of digital digital in technologies such as Ethernet, USB and PCI Express. This allows the use of internal and external display connections, and unlike outdated standards that transmit an hour-long signal with each output, the DisplayPort protocol is based on small data packets known as micro-packs that can insert a clock signal into the data stream. This allows you to get a higher resolution using fewer contacts. Using data packets also makes DisplayPort possible to be exuding, which means that additional features can be added over time without significant changes to the physical interface. DisplayPort can be used to transmit audio and video at the same time, although each one is optional and can be streamed without the other. The path of the video signal can range from six to sixteen bits per color channel, and the audio path can have up to eight channels of 24-bit, 192 kHz PCM audio that is not compressed. A two-direction, semi-duplex support channel carries device management and device management data for the main link, such as VESA EDID, MCCS, and DPMS standards. In addition, the interface is capable of carrying two-direction USB signals. DisplayPort uses an LVDS signal protocol that is not compatible with DVI or HDMI. However, dual-mode Ports DisplayPort are designed to transmit the DVI or HDMI protocol with a single link (TMDS) through the interface using an external passive adapter. This adapter provides compatibility mode and converts the signal from 3.3 volts to 5 volts. Analog VGA/YPbPr and two-edge DVI require an active powered adapter for compatibility and does not rely on dual mode. Active VGA adapters are powered directly by the DisplayPort connector, while active dual-link DVI adapters typically rely on an external power source such as USB. Versions 1.0 to 1.1 The first version, 1.0, was approved by VESA on May 3, 2006. Version 1.1 was ratified on April 2, 2007, and version 1.1a was ratified on January 11, 2008. DisplayPort 1.0-1.1a allows a maximum bandwidth of 10.8 Gbit/s (8.64 Gbit/s data speed) over the standard 4-band main link. DisplayPort cables up to 2 metres long are required to maintain a full capacity of 10.8 Gbps. DisplayPort 1.1 allows devices to implement alternative layers of communication, such as fiber optic, allowing for much longer reach between the source and the display without signal degradation, although alternative implementations are not standardized. It also includes HDCP in addition to DisplayPort Content Protection (DPCP). DisplayPort 1.1a is free to download from the VESA website. Version 1.2 DisplayPort 1.2 was introduced on January 7, 2010. The most significant improvement in the new version is the doubling of bandwidth up to 17.28 Gbps in High Bit Rate 2 (HBR2) mode, allowing you to increase resolution, higher upgrade speeds and greater color depth. Other improvements include yourself independent video streams (multi-stream transport) called Multi-Stream Transport, stereoscopic 3D tools, AUX bandwidth increase (from 1 Mbps to 720 Mbps), more color spaces including xvYCC, scRGB and Adobe RGB 1998, and Global Time Code (GTC) to synchronize audio/video. In addition, Apple Inc.'s Mini DisplayPort connector, which is much smaller and is designed for laptops and other smaller devices, is compatible with the new standard. Version 1.2a DisplayPort 1.2a was released in January 2013 and may additionally include adaptive VESA synchronization. AmD FreeSync uses DisplayPort Adaptive-Sync to work. FreeSync was first demonstrated at CES 2014 on the Toshiba Satellite laptop, using the Panel-Self-Refresh (PSR) feature from Embedded DisplayPort, and after an offer from AMD, VESA later adapted the Panel-Self-Refresh feature for use in standalone displays and added it as an additional feature of the main DisplayPort standard called Adaptive-Sync Because it is a non-binding feature, Adaptive-Sync support is not required for display to be DisplayPort 1.2a compatible. 1.3 Version 1.3 DisplayPort was approved on September 15, 2014. This standard increases the total transmission capacity to 32.4 Gbps with the new HBR3 mode from 8.1 Gbps per lane (according to 5.4 Gbps HBR2 in version 1.2), for total data bandwidth of 25.92 Gbit/s after factoring in 8b/10b encoding overheads. This bandwidth is enough for a 4K UHD display (3840 × 2160) at 120 Hz with 24 bits / px RGB colors, 5K display (5120 × 2880) at 60 Hz with 30 bits / px RGB color, or 8K UHD display (7680 × 4320) at 30Hz with 24 bits / px RGB color. Using Multi-Stream Transport (MST), DisplayPort can control two 4K UHD (3840 × 2160) displays at 60 Hz or up to four W'HXGA displays (2560 × 1600) at 60 Hz with a 24-bit/px RGB color. The new standard includes mandatory dual mode for DVI and HDMI adapters, introduction of HDMI 2.0 and HDCP 2.2 content protection. The Thunderbolt 3 connection standard was originally supposed to include DisplayPort 1.3, but the final release ended with only version 1.2. VeSA adaptive synchronization in 1.3 DisplayPort remains an additional part of the specification. Version 1.4 displayPort was published on March 1, 2016. The new transmission modes are not defined, so HBR3 (32.4 Gbit/s) presented in version 1.3 is still the most affordable mode. DisplayPort 1.4 adds support for compression display flow 1.2 (DSC), Correction errors forward, HDR10 metadata defined in CTA-861.3, including static and dynamic metadata and Rec color space. 2020, for HDMI compatibility, and maximum number of stationary audio channels of audio channels DSC is a visually no loss method of coding with a compression ratio of up to 3:1. Using HBR3 DSC, DisplayPort 1.4 can support 8K UHD (7680 × 4320) at 60Hz or 4K UHD (3840 × 2160) at 120 Hz with 30 bit/px RGB and HDR. 4K by 60 Hz 30 bits / px RGB/ HDR can be achieved without the need for DSC. On displays that do not support DSC, maximum limits remain unchanged from DisplayPort 1.3 (4K 120 Hz, 5K 60 Hz, 8K 30Hz). Version 1.4a DisplayPort 1.4a was published in April 2018. VESA has not made an official press release for this version. It updated the DSC DisplayPort implementation from DSC 1.2 to 1.2a. According to a roadmap published by VESA in September 2016, the new version of DisplayPort was to be launched in early 2017. This would improve the communication speed from 8.1 to 10.0 Gbps, which is 24% more.
Recommended publications
  • Scalable Multi-View Stereo Camera Array for Real World Real-Time Image Capture and Three-Dimensional Displays
    Scalable Multi-view Stereo Camera Array for Real World Real-Time Image Capture and Three-Dimensional Displays Samuel L. Hill B.S. Imaging and Photographic Technology Rochester Institute of Technology, 2000 M.S. Optical Sciences University of Arizona, 2002 Submitted to the Program in Media Arts and Sciences, School of Architecture and Planning in Partial Fulfillment of the Requirements for the Degree of Master of Science in Media Arts and Sciences at the Massachusetts Institute of Technology June 2004 © 2004 Massachusetts Institute of Technology. All Rights Reserved. Signature of Author:<_- Samuel L. Hill Program irlg edia Arts and Sciences May 2004 Certified by: / Dr. V. Michael Bove Jr. Principal Research Scientist Program in Media Arts and Sciences ZA Thesis Supervisor Accepted by: Andrew Lippman Chairperson Department Committee on Graduate Students MASSACHUSETTS INSTITUTE OF TECHNOLOGY Program in Media Arts and Sciences JUN 172 ROTCH LIBRARIES Scalable Multi-view Stereo Camera Array for Real World Real-Time Image Capture and Three-Dimensional Displays Samuel L. Hill Submitted to the Program in Media Arts and Sciences School of Architecture and Planning on May 7, 2004 in Partial Fulfillment of the Requirements for the Degree of Master of Science in Media Arts and Sciences Abstract The number of three-dimensional displays available is escalating and yet the capturing devices for multiple view content are focused on either single camera precision rigs that are limited to stationary objects or the use of synthetically created animations. In this work we will use the existence of inexpensive digital CMOS cameras to explore a multi- image capture paradigm and the gathering of real world real-time data of active and static scenes.
    [Show full text]
  • 3D Techniques
    3D Techniques Univ.Prof. Dr.-Ing. Markus Rupp LVA 389.141 Fachvertiefung Telekommunikation (LVA: 389.137 Image and Video Compression) Last change: Jan 20, 2020 Outline • Binocluar Vision • Stereo Images: From first approaches to standardisation • 3D TV Univ.-Prof. Dr.-Ing. Markus Rupp 2 So far we did only this… This is an eye-pad not an IPad Binocular vision • Binocular vision is vision in which both eyes are used together. The word binocular comes from two Latin roots, bini for double, and oculus for eye.[1] Having two eyes confers at least four advantages over having one. First, it gives a creature a spare eye in case one is damaged. Second, it gives a wider field of view. For example, humans have a maximum horizontal field of view of approximately 200 degrees with two eyes, approximately 120 degrees of which makes up the binocular field of view (seen by both eyes) flanked by two uniocular fields (seen by only one eye) of approximately 40 degrees. [2] Third, it gives binocular summation in which the ability to detect faint objects is enhanced.[3] Fourth it can give stereopsis in which parallax provided by the two eyes' different positions on the head give precise depth perception.[4] Such binocular vision is usually accompanied by singleness of vision or binocular fusion, in which a single image is seen despite each eye's having its own image of any object.[4] Binocular vision • Some animals, usually prey animals, have their two eyes positioned on opposite sides of their heads to give the widest possible field of view.
    [Show full text]
  • Dr. Juan Carlos Jácome Fernández
    Title CAPTURING OF 3D CONTENT USING A SINGLE APERTURE CAMERA Name: Dr. Juan Carlos Jácome Fernández This is a digitised version of a dissertation submitted to the University of Bedfordshire. It is available to view only. This item is subject to copyright. CAPTURING OF 3D CONTENT USING A SINGLE APERTURE CAMERA By Juan Carlos Jácome Fernández A thesis submitted to the University of Bedfordshire in partial fulfilment of the requirements for the degree of Doctor of Philosophy October 2017 Abstract Integral imaging has recently re-emerged as an alternative to current 3D capturing systems, providing post-production refocusing capabilities and reducing the complexity of 3D capturing systems. One of the main drawbacks of conventional plenoptic 1 integral imaging systems is the implementation of a single custom made microlens array which has a fixed focal length and a high cost/low scalability associated with. This thesis demonstrates a variable focal length microlens arrays system, which can flexibly operate within a range of various focal lengths, increase the cost-effectiveness of the integral imaging system and offers the opportunity to manipulate the main camera optical parameters without modifying the objective lens settings. To validate the proposed concept, a custom-made integral imaging camera system was designed and built (IMPERX 4K system). Based on the results obtained from two initial optical simulations, software simulation and mathematical model; the suitable microlens arrays were acquired, and several experiments were performed to establish the feasibility of a variable focal length microlens arrays system. The results obtained show the proposed system performed as expected. The variable focal length system represents an ideal method to control microlens focal length, as it offers a higher microlens count, without a dedicated electronic system, whose performance is not susceptible to temperature-pressure changes and can operate in real-time as it does not require downtimes to re-adjust.
    [Show full text]
  • Proximal Monitoring in Landscape Environments
    Department of Mathematics and Statistics Proximal Monitoring in Landscape Environments Shuqi Ng This dissertation is presented for the Degree of Doctor of Philosophy of Curtin University December 2017 DECLARATION To the best of my knowledge and belief, this thesis contains no material previously published by any other person except where due acknowledgement has been made. This thesis contains no material which has been accepted for the award of any other degree or diploma in any university. Shuqi Ng March 2017 ABSTRACT With the implementation of the initiatives in reducing emissions from deforestation and forest degradation (REDD), the accurate determination of the spatio-temporal variation of carbon stocks is crucial. Woody vegetation is one of the more noteworthy carbon storage pools. However, ever-changing forest inventory makes it difficult for countries to accurately measure woody biomass and by extension, predict carbon stocks. Although the most accurate mensuration of biomass is to harvest a tree, the method is destructive and obstructs the REDD initiative. Non- destructive methods use dendrometric measurements that have been obtained from non-contact, remote sensing technologies to estimate biomass derived from allometric models. The commodity stereo-camera is an alternate consideration to remote sensing technologies currently used such as terrestrial laser scanning (TLS), which requires specialized equipment and some expertise in using the equipment. Improving technologies and software has enabled the photogrammetric measurements of objects to be reasonably accurate comparative to TLS. As a more mobile and cost-effective equipment, it also addresses some of the issues in using laser technology. In photogrammetric measurement, the application of spatial scale to the model is critical for accurate distance and volume estimates (Miller (2015)).
    [Show full text]
  • Synthesis, Coding, and Evaluation of 3D Images Based on Integral Imaging
    Synthesis, Coding, and Evaluation of 3D Images Based on Integral Imaging Roger Olsson Department of Information Technology and Media Mid Sweden University Doctoral Thesis No. 55 Sundsvall, Sweden 2008 Mittuniversitetet Informationsteknologi och medier ISBN 978-91-85317-98-1 SE-851 70 Sundsvall ISSN 1652-893X SWEDEN Akademisk avhandling som med tillstånd av Mittuniversitetet framlägges till offent- lig granskning för avläggande av Teknologie Doktorsexamen torsdagen den 12 juni 2008 i L111, Mittuniversitetet, Holmgatan 10, Sundsvall. c Roger Olsson, maj 2008 Tryck: Kopieringen, Mittuniversitetet, Sundsvall. Always in motion is the future Yoda iv Sammanfattning De senaste åren har kameraprototyper som kan fånga tredimensionella (3D) bilder presenterats, baserade på 3D-tekniken Integral Imaging (II). När dessa II-bilder be- traktas på en 3D-skärm, delger de både ett djup och ett innehåll som på ett realistiskt sätt ändrar perspektiv när tittaren ändrar sin betraktningsposition. Avhandlingen koncentrerar sig på tre hämmande faktorer gällande II-bilder. För det första finns det en mycket begränsad allmän tillgång till II-bilder för jämförande forskning och utveckling av kodningsmetoder. Det finns heller inga objektiva kvali- tetsmått som uttryckligen mäter distorsion med avseende på II-bildens egenskaper: djup och betraktningsvinkelberoende. Slutligen uppnår nuvarande standarder för bildkodning låg kodningseffektivitet när de appliceras på II-bilder. En metod baserad på datorrendrering har utvecklats som tillåter produktion av olika typer av II-bilder. En II-kameramodell ingår som bas, kombinerad med ett scen- beskrivningsspråk som möjligör att godtydligt komplexa virtuella scener definieras. Ljustransporten inom scenen och fram till II-kameran simuleras med strålföljning och geometrisk optik. Den presenterade metoden används för att skapa ett antal II- kameramodeller, scendefinitioner och II-bilder.
    [Show full text]
  • Generation of 3-D Synthetic Autostereoscopic Integral Images Using Computer Simulated Cameras
    GENERATION OF 3-D SYNTHETIC AUTOSTEREOSCOPIC INTEGRAL IMAGES USING COMPUTER SIMULATED CAMERAS A thesis submitted in partial fulfilment of the requirements for the degree of Doctor of Philosophy By Shafik Salih Electronic and Computer Engineering School of Engineering and Design Brunel University March 2014 Abstract ________________________________________________________ Production of artificial Three-Dimension (3-D) images was the aim of many researches over hundreds of years. 3-D images are the images that create sense of depth when viewing them. 3-D images are closer to the real world scenes than 2-D images due to the 3-D effect or the sense of depth the 3-D images provide. Sense of depth can be caused by binocular cues including convergence and parallax. Convergence is created by the difference between the angles of the left eye and the right eye viewing axes. Parallax is the effect of viewing with one eye a view of the scene that is inherently shifted to the view seen by the other eye. Several techniques have targeted the creation of 3-D images with the mentioned cues. The technique is preferred when it is able to create 3-D images so that the viewer can view these images without wearing special glasses and the occurrence of viewer fatigue. Integral photography that was invented in 1908 is able to meet the previous requirements. Based on integral photography, several techniques, research and studies have been published. The purposes of this thesis include the computer simulation of flexible integral photography systems, the computer generation of good quality 3-D static and animated integral images using the simulated systems, optimising the generation process to be more accurate, less expensive, more effective, and faster, and producing a portable specialist software tool to achieve these targets.
    [Show full text]
  • A Method for the Assessment of the Quality of Stereoscopic 3D Images Raluca Vlad
    A Method for the Assessment of the Quality of Stereoscopic 3D Images Raluca Vlad To cite this version: Raluca Vlad. A Method for the Assessment of the Quality of Stereoscopic 3D Images. Signal and Image Processing. Université de Grenoble, 2013. English. tel-00925280v2 HAL Id: tel-00925280 https://tel.archives-ouvertes.fr/tel-00925280v2 Submitted on 10 Jan 2014 (v2), last revised 28 Feb 2014 (v3) HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. THÈSE pour obtenir le grade de DOCTEUR DE L’UNIVERSITÉ DE GRENOBLE Spécialité : Signal, Image, Parole, Télécommunications (SIPT) Arrêté ministériel : 7 août 2006 Présentée par Mlle Raluca VLAD Thèse dirigée par Mme Anne GUÉRIN-DUGUÉ et Mme Patricia LADRET préparée au sein du laboratoire Grenoble - Images, Parole, Signal, Automatique (GIPSA-lab) dans l’école doctorale d’Électronique, Électrotechnique, Automatique et Traitement du signal (EEATS) Une méthode pour l’évaluation de la qualité des images 3D stéréoscopiques Thèse soutenue publiquement le 2 décembre 2013, devant le jury composé de : Patrick LE CALLET
    [Show full text]
  • Stereo World Magazine and Enroli Me As a Bad Time to Invest Not in Just Rare Member of the National Stereoscopic Association
    3-0 Imaging Past & Present November/December 2008 Volume 34, Number 3 A taste of tho lato qos through tho oady '608 found in amatour storm slidas bg~uk~lll~c I were made. Thev are both mount- watching the adventures of the Space Gear for Christmas ed in the old st);le gray Kodak Robinson family in Lost in Space, t's not often that both images in heat-seal mounts with red edges. which I'm sure had numerous tie- this column are of the same sub- These two brothers apparently in products that made their way to ject, taken within moments of got up CMstmas morning and CMstmas wish lists, but that was I found space helmets (complete probably nearly a decade after each other, but this pair just seemed too good not to share with inflatable shoulder pads) and these guys became space explorers. them both! weapons waiting for them under The photographer must have These slides were found in an the tree. Since this was a bit before been using a rather slow exposure accumulation of images that my time, I don't know if these out- setting and moved the camera fits were offidal replicas based on when taking these photos, since appear to have been taken in the TV 1950s, although these particular some show or comic book flames from the fireplace can be ones contain no notes or informa- series, or if they were just generic seen through the one boy's pants, when space toys. Personally, I grew up and I'm sure he's not on fire! tion as to or where they Fortunately for the boys, these bubble-style helmets appear to have had large rectangular open- ings in front of their faces, which would not be so useful in space, but would have kept young explor- ers on earth supplied with enough oxygen to avoid problems.
    [Show full text]
  • Atelier 3D 20/04/2011
    Clubs IBM Photo + Video + Micro Atelier 3D 20/04/2011 ChristopheChristophe DentingerDentinger Sommaire de l’Atelier 3D Partie ‘théorique’ (40 minutes) Origines / Fondement de la 3D / Stéréoscopie / Histoire des techniques 3D Techniques et matériels prises de vue Photo et Vidéo 3D Techniques et matériels de Visualisation 3D / Logiciels Ateliers parallèles (3 x 20 minutes en 3 groupes) Salle Audiovisuelle: Démonstration de l’adaptateur réflex 3D Loreo, Démonstration du Fuji W3, Visualisation photos avec Loreo Pixi 3D Viewer Club Video: Démonstration GoPro, Réglage effet 3D sous Cineform studio, Exportation des vidéos et visualisation au format Anaglyphe rouge / cyan Club Micro: Visualisation avec lunettes actives NVIDIA de photos et vidéos prises avec le Fuji W3 et L’adaptateur Loreo 3D Utilisation de Myfinepix studio pour gérer les fichiers 3D MPO / AVI Utilisation de Stereo Photo Maker pour aligner et exporter les photos 3D Page 2 Origines / Fondements de la 3D Fondements: La 3D est liée principalement à l’un de nos 5 sens (la vision) et l’interaction avec l’espace qui nous entoure Espace et 3D ‘Naturelle’ L’espace est une notion qui correspond à la perception de notre environnement naturel physique par l’un de nos sens (vue). L’espace géométrique euclidien est une représentation mathématique qui modélise cet environnement (par exemple 2 dimensions pour le sol = 4 directions et la 3ième dimension pour la hauteur), l’axe vertical étant ‘causé’ par la gravité. La part de l’évolution dans la vision Adaptation des espèces animales: Par exemple les animaux chassés ont une vision panoramique alors que les prédateurs doivent percevoir les distances.
    [Show full text]
  • A Uav-Based Low-Cost Stereo Camera System for Archaeological Surveys - Experiences from Doliche (Turkey)
    International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences, Volume XL-1/W2, 2013 UAV-g2013, 4 – 6 September 2013, Rostock, Germany A UAV-BASED LOW-COST STEREO CAMERA SYSTEM FOR ARCHAEOLOGICAL SURVEYS - EXPERIENCES FROM DOLICHE (TURKEY) K. Haubeck a, T. Prinz b a Institute of Geography, Westfalian Wilhelms-University Münster, Germany – [email protected] b Institute for Geoinformatics, Westfalian Wilhelms-University Münster, Germany – [email protected] KEY WORDS: Archaeology, Application, DEM/DTM, Orthoimage, Close Range, UAV, Stereoscopic ABSTRACT: The use of Unmanned Aerial Vehicles (UAVs) for surveying archaeological sites is becoming more and more common due to their advantages in rapidity of data acquisition, cost-efficiency and flexibility. One possible usage is the documentation and visualization of historic geo-structures and -objects using UAV-attached digital small frame cameras. These monoscopic cameras offer the possibility to obtain close-range aerial photographs, but – under the condition that an accurate nadir-waypoint flight is not possible due to choppy or windy weather conditions – at the same time implicate the problem that two single aerial images not always meet the required overlap to use them for 3D photogrammetric purposes. In this paper, we present an attempt to replace the monoscopic camera with a calibrated low-cost stereo camera that takes two pictures from a slightly different angle at the same time. Our results show that such a geometrically predefined stereo image pair can be used for photogrammetric purposes e.g. the creation of digital terrain models (DTMs) and orthophotos or the 3D extraction of single geo-objects.
    [Show full text]
  • The 3D New Wave and the Future Image ~ Trend of 3D Autostereoscopic Display and Its Future Image ~
    Fujiwara-Rothchild, Ltd. Trend of 3D autostereoscopic display and its future image July, 2010 Market Research Report published on July, 2010 The 3D new wave and the future image ~ Trend of 3D autostereoscopic display and Its future image ~ Fujiwara-Rothchild, Ltd. Y’s Bldg. 3F, 3-6-15 Kanda-Jimbocho, Chiyoda-ku, Tokyo 101-0051 JAPAN Phone : 03-3239-3008 FAX: 03-3239-8081 E-mail: [email protected] The 3D new wave and the future image Copyright 2010 Fujiwara-Rothchild, Ltd. All Rights Reserved. Y’s Bldg. 3F, 3-6-15 Kanda-Jimbocho, Chiyoda-ku, Tokyo 101-0051 JAPAN Phone:81-3-3239-3008 Fax:81-3-3239-8081 Email:[email protected] 1 Fujiwara-Rothchild, Ltd. Trend of 3D autostereoscopic display and its future image July, 2010 INDEX 1. Preface ........................................................................................................ 7 2. Exective Summary ..................................................................................... 9 3. 3D back groung ........................................................................................ 12 3.1. 3D History ........................................................................................................... 12 3.2. Trend of Film industry ........................................................................................ 13 3.2.1. Movie business ............................................................................................. 13 3.2.2. Active demand of recent Hollywood 3D movie ........................................... 15 3.2.3. 3D Movie bussiness .....................................................................................
    [Show full text]
  • Image and Video Compression Coding Theory Contents
    Image and Video Compression Coding Theory Contents 1 JPEG 1 1.1 The JPEG standard .......................................... 1 1.2 Typical usage ............................................. 1 1.3 JPEG compression ........................................... 2 1.3.1 Lossless editing ........................................ 2 1.4 JPEG files ............................................... 3 1.4.1 JPEG filename extensions ................................... 3 1.4.2 Color profile .......................................... 3 1.5 Syntax and structure .......................................... 3 1.6 JPEG codec example ......................................... 4 1.6.1 Encoding ........................................... 4 1.6.2 Compression ratio and artifacts ................................ 8 1.6.3 Decoding ........................................... 10 1.6.4 Required precision ...................................... 11 1.7 Effects of JPEG compression ..................................... 11 1.7.1 Sample photographs ...................................... 11 1.8 Lossless further compression ..................................... 11 1.9 Derived formats for stereoscopic 3D ................................. 12 1.9.1 JPEG Stereoscopic ...................................... 12 1.9.2 JPEG Multi-Picture Format .................................. 12 1.10 Patent issues .............................................. 12 1.11 Implementations ............................................ 13 1.12 See also ................................................ 13 1.13 References
    [Show full text]