Gstreamer and Dmabuf
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
GLSL 4.50 Spec
The OpenGL® Shading Language Language Version: 4.50 Document Revision: 7 09-May-2017 Editor: John Kessenich, Google Version 1.1 Authors: John Kessenich, Dave Baldwin, Randi Rost Copyright (c) 2008-2017 The Khronos Group Inc. All Rights Reserved. This specification is protected by copyright laws and contains material proprietary to the Khronos Group, Inc. It or any components may not be reproduced, republished, distributed, transmitted, displayed, broadcast, or otherwise exploited in any manner without the express prior written permission of Khronos Group. You may use this specification for implementing the functionality therein, without altering or removing any trademark, copyright or other notice from the specification, but the receipt or possession of this specification does not convey any rights to reproduce, disclose, or distribute its contents, or to manufacture, use, or sell anything that it may describe, in whole or in part. Khronos Group grants express permission to any current Promoter, Contributor or Adopter member of Khronos to copy and redistribute UNMODIFIED versions of this specification in any fashion, provided that NO CHARGE is made for the specification and the latest available update of the specification for any version of the API is used whenever possible. Such distributed specification may be reformatted AS LONG AS the contents of the specification are not changed in any way. The specification may be incorporated into a product that is sold as long as such product includes significant independent work developed by the seller. A link to the current version of this specification on the Khronos Group website should be included whenever possible with specification distributions. -
Report of Contributions
X.Org Developers Conference 2020 Report of Contributions https://xdc2020.x.org/e/XDC2020 X.Org Developer … / Report of Contributions State of text input on Wayland Contribution ID: 1 Type: not specified State of text input on Wayland Wednesday, 16 September 2020 20:15 (5 minutes) Between the last impromptu talk at GUADEC 2018, text input on Wayland has become more organized and more widely adopted. As before, the three-pronged approach of text_input, in- put_method, and virtual keyboard still causes confusion, but increased interest in implementing it helps find problems and come closer to something that really works for many usecases. The talk will mention how a broken assumption causes a broken protocol, and why we’re notdone with Wayland input methods yet. It’s recommended to people who want to know more about the current state of input methods on Wayland. Recommended background: aforementioned GUADEC talk, wayland-protocols reposi- tory, my blog: https://dcz_self.gitlab.io/ Code of Conduct Yes GSoC, EVoC or Outreachy No Primary author: DCZ, Dorota Session Classification: Demos / Lightning talks I Track Classification: Lightning Talk September 30, 2021 Page 1 X.Org Developer … / Report of Contributions IGT GPU Tools 2020 Update Contribution ID: 2 Type: not specified IGT GPU Tools 2020 Update Wednesday, 16 September 2020 20:00 (5 minutes) Short update on IGT - what has changed in the last year, where are we right now and what we have planned for the near future. IGT GPU Tools is a collection of tools and tests aiding development of DRM drivers. It’s widely used by Intel in its public CI system. -
Blackberry QNX Multimedia Suite
PRODUCT BRIEF QNX Multimedia Suite The QNX Multimedia Suite is a comprehensive collection of media technology that has evolved over the years to keep pace with the latest media requirements of current-day embedded systems. Proven in tens of millions of automotive infotainment head units, the suite enables media-rich, high-quality playback, encoding and streaming of audio and video content. The multimedia suite comprises a modular, highly-scalable architecture that enables building high value, customized solutions that range from simple media players to networked systems in the car. The suite is optimized to leverage system-on-chip (SoC) video acceleration, in addition to supporting OpenMAX AL, an industry open standard API for application-level access to a device’s audio, video and imaging capabilities. Overview Consumer’s demand for multimedia has fueled an anywhere- o QNX SDK for Smartphone Connectivity (with support for Apple anytime paradigm, making multimedia ubiquitous in embedded CarPlay and Android Auto) systems. More and more embedded applications have require- o Qt distributions for QNX SDP 7 ments for audio, video and communication processing capabilities. For example, an infotainment system’s media player enables o QNX CAR Platform for Infotainment playback of content, stored either on-board or accessed from an • Support for a variety of external media stores external drive, mobile device or streamed over IP via a browser. Increasingly, these systems also have streaming requirements for Features at a Glance distributing content across a network, for instance from a head Multimedia Playback unit to the digital instrument cluster or rear seat entertainment units. Multimedia is also becoming pervasive in other markets, • Software-based audio CODECs such as medical, industrial, and whitegoods where user interfaces • Hardware accelerated video CODECs are increasingly providing users with a rich media experience. -
The Opencl Specification
The OpenCL Specification Version: 2.0 Document Revision: 22 Khronos OpenCL Working Group Editor: Aaftab Munshi Last Revision Date: 3/18/14 Page 1 1. INTRODUCTION ............................................................................................................... 10 2. GLOSSARY ......................................................................................................................... 12 3. THE OPENCL ARCHITECTURE .................................................................................... 23 3.1 Platform Model ................................................................................................................................ 23 3.2 Execution Model .............................................................................................................................. 25 3.2.1 Execution Model: Mapping work-items onto an NDRange ........................................................................28 3.2.2 Execution Model: Execution of kernel-instances ........................................................................................30 3.2.3 Execution Model: Device-side enqueue ......................................................................................................31 3.2.4 Execution Model: Synchronization .............................................................................................................32 3.2.5 Execution Model: Categories of Kernels ....................................................................................................33 3.3 Memory -
Openbricks Embedded Linux Framework - User Manual I
OpenBricks Embedded Linux Framework - User Manual i OpenBricks Embedded Linux Framework - User Manual OpenBricks Embedded Linux Framework - User Manual ii Contents 1 OpenBricks Introduction 1 1.1 What is it ?......................................................1 1.2 Who is it for ?.....................................................1 1.3 Which hardware is supported ?............................................1 1.4 What does the software offer ?............................................1 1.5 Who’s using it ?....................................................1 2 List of supported features 2 2.1 Key Features.....................................................2 2.2 Applicative Toolkits..................................................2 2.3 Graphic Extensions..................................................2 2.4 Video Extensions...................................................3 2.5 Audio Extensions...................................................3 2.6 Media Players.....................................................3 2.7 Key Audio/Video Profiles...............................................3 2.8 Networking Features.................................................3 2.9 Supported Filesystems................................................4 2.10 Toolchain Features..................................................4 3 OpenBricks Supported Platforms 5 3.1 Supported Hardware Architectures..........................................5 3.2 Available Platforms..................................................5 3.3 Certified Platforms..................................................7 -
Augmented Reality Navigation
VYSOKÉ UČENÍ TECHNICKÉ V BRNĚ BRNO UNIVERSITY OF TECHNOLOGY FAKULTA ELEKTROTECHNIKY A KOMUNIKAČNÍCH TECHNOLOGIÍ ÚSTAV RADIOELEKTRONIKY FACULTY OF ELECTRICAL ENGINEERING AND COMMUNICATION DEPARTMENT OF RADIO ELECTRONICS AUGMENTED REALITY APPLICATIONS IN EMBEDDED NAVIGATION DEVICES BAKALÁŘSKÁ PRÁCE BACHELOR'S THESIS AUTOR PRÁCE MARTIN JAROŠ AUTHOR BRNO 2014 VYSOKÉ UČENÍ TECHNICKÉ V BRNĚ BRNO UNIVERSITY OF TECHNOLOGY FAKULTA ELEKTROTECHNIKY A KOMUNIKAČNÍCH TECHNOLOGIÍ ÚSTAV RADIOELEKTRONIKY FACULTY OF ELECTRICAL ENGINEERING AND COMMUNICATION DEPARTMENT OF RADIO ELECTRONICS AUGMENTED REALITY APPLICATIONS IN EMBEDDED NAVIGATION DEVICES AUGMENTED REALITY APPLICATIONS IN EMBEDDED NAVIGATION DEVICES BAKALÁŘSKÁ PRÁCE BACHELOR'S THESIS AUTOR PRÁCE MARTIN JAROŠ AUTHOR VEDOUCÍ PRÁCE doc. Ing. TOMÁŠ FRÝZA, Ph.D. SUPERVISOR BRNO 2014 VYSOKÉ UČENÍ TECHNICKÉ V BRNĚ Fakulta elektrotechniky a komunikačních technologií Ústav radioelektroniky Bakalářská práce bakalářský studijní obor Elektronika a sdělovací technika Student: Martin Jaroš ID: 146847 Ročník: 3 Akademický rok: 2013/2014 NÁZEV TÉMATU: Augmented reality applications in embedded navigation devices POKYNY PRO VYPRACOVÁNÍ: Analyze the hardware possibilities of the OMAP platform and design an application to effectively combine captured video data and rendered virtual scene based on navigational data from GPS and INS sensors. Design and create a functional prototype. Examine practical use cases of the proposed navigation device, design applicable user interface. DOPORUČENÁ LITERATURA: [1] BIMBER, O.; RASKAR, R. Spatial augmented reality: merging real and virtual worlds. Wellesley: A K Peters, 2005, 369 p. ISBN 15-688-1230-2. [2] Texas Instruments. OMAP 4460 Multimedia Device [online]. 2012 - [cit. 8. listopadu 2012]. Available: http://www.ti.com/product/omap4460. Termín zadání: 10.2.2014 Termín odevzdání: 30.5.2014 Vedoucí práce: doc. -
The Openvx™ Specification
The OpenVX™ Specification Version 1.2 Document Revision: dba1aa3 Generated on Wed Oct 11 2017 20:00:10 Khronos Vision Working Group Editor: Stephen Ramm Copyright ©2016-2017 The Khronos Group Inc. i Copyright ©2016-2017 The Khronos Group Inc. All Rights Reserved. This specification is protected by copyright laws and contains material proprietary to the Khronos Group, Inc. It or any components may not be reproduced, republished, distributed, transmitted, displayed, broadcast or otherwise exploited in any manner without the express prior written permission of Khronos Group. You may use this specifica- tion for implementing the functionality therein, without altering or removing any trademark, copyright or other notice from the specification, but the receipt or possession of this specification does not convey any rights to reproduce, disclose, or distribute its contents, or to manufacture, use, or sell anything that it may describe, in whole or in part. Khronos Group grants express permission to any current Promoter, Contributor or Adopter member of Khronos to copy and redistribute UNMODIFIED versions of this specification in any fashion, provided that NO CHARGE is made for the specification and the latest available update of the specification for any version of the API is used whenever possible. Such distributed specification may be re-formatted AS LONG AS the contents of the specifi- cation are not changed in any way. The specification may be incorporated into a product that is sold as long as such product includes significant independent work developed by the seller. A link to the current version of this specification on the Khronos Group web-site should be included whenever possible with specification distributions. -
The Road to the Mainline Zynqmp VCU Driver
The Road to the Mainline ZynqMP VCU Driver FOSDEM ’21 Michael Tretter – [email protected] https://www.pengutronix.de Agenda Xilinx Zynq® UltraScale+™ MPSoC H.264/H.265 Video Codec Unit Video Encoders in Mainline Linux VCU Mainline Driver: Allegro A Glimpse into the Future 2/47 Xilinx Zynq® UltraScale+™ MPSoC 3/47 ZynqMP Platform Overview Luca Ceresoli: ARM64 + FPGA and more: Linux on the Xilinx ZynqMP https://archive.fosdem.org/ 2018/schedule/event/arm6 4_and_fpga 4/47 ZynqMP Mainline Status Mainline Linux just works, e.g., on ZCU104 Evaluation Kit U-Boot, Barebox, FSBL Sometimes more reliable with Xilinx downstream Xilinx is actively mainlining their drivers 5/47 Make Sure that Your ZynqMP has a VCU ZU # E V ZU: Zynq Ultrascale+ #: Value Index C/E: Processor System Identifier G/V: Engine Type 6/47 Focus on Video Encoding VCU supports video decoding, as well Linux mainline driver only supports encoding Decoding might be focus in a future talk 7/47 Basic Video Encoding Knowledge Expected Paul Kocialkowski: Supporting Hardware-Accelerated Video Encoding with Mainline https://www.youtube.com/watch?v=S5wCdZfGFew 8/47 H.264/H.265 Video Codec Unit 9/47 VCU: Documentation Hardware configuration Software usage Available on the Xilinx Website 10/47 VCU: Features The encoder engine is designed to process video streams using the HEVC (ISO/IEC 23008-2 high-efficiency Video Coding) and AVC (ISO/IEC 14496-10 Advanced Video Coding) standards. It provides complete support for these standards, including support for 8-bit and 10-bit color, Y- only (monochrome), 4:2:0 and 4:2:2 Chroma formats, up to 4K UHD at 60 Hz performance. -
The Opengl ES Shading Language
The OpenGL ES® Shading Language Language Version: 3.20 Document Revision: 12 246 JuneAugust 2015 Editor: Robert J. Simpson, Qualcomm OpenGL GLSL editor: John Kessenich, LunarG GLSL version 1.1 Authors: John Kessenich, Dave Baldwin, Randi Rost 1 Copyright (c) 2013-2015 The Khronos Group Inc. All Rights Reserved. This specification is protected by copyright laws and contains material proprietary to the Khronos Group, Inc. It or any components may not be reproduced, republished, distributed, transmitted, displayed, broadcast, or otherwise exploited in any manner without the express prior written permission of Khronos Group. You may use this specification for implementing the functionality therein, without altering or removing any trademark, copyright or other notice from the specification, but the receipt or possession of this specification does not convey any rights to reproduce, disclose, or distribute its contents, or to manufacture, use, or sell anything that it may describe, in whole or in part. Khronos Group grants express permission to any current Promoter, Contributor or Adopter member of Khronos to copy and redistribute UNMODIFIED versions of this specification in any fashion, provided that NO CHARGE is made for the specification and the latest available update of the specification for any version of the API is used whenever possible. Such distributed specification may be reformatted AS LONG AS the contents of the specification are not changed in any way. The specification may be incorporated into a product that is sold as long as such product includes significant independent work developed by the seller. A link to the current version of this specification on the Khronos Group website should be included whenever possible with specification distributions. -
Field Service Support with Google Glass and Webrtc
Field Service Support with Google Glass and WebRTC Support av Fälttekniker med Google Glass och WebRTC PATRIK OLDSBERG Examensarbete inom Datorteknik, Grundnivå, 15 hp Handledare på KTH: Reine Bergström Examinator: Ibrahim Orhan TRITA-STH 2014:68 KTH Skolan för Teknik och Hälsa 136 40 Handen, Sverige Abstract The Internet is dramatically changing the way we com- municate, and it is becoming increasingly important for communication services to adapt to context in which they are used. The goal of this thesis was to research how Google Glass and WebRTC can be used to create a communi- cation system tailored for field service support. A prototype was created where an expert is able to provide guidance for a field technician who is wearing Google Glass. A live video feed is sent from Glass to the expert from which the expert can select individual images. When a still image is selected it is displayed to the technician through Glass, and the expert is able to provide instructions using real time annotations. An algorithm that divides the selected image into segments was implemented using WebGL. This made it possible for the expert to highlight objects in the image by clicking on them. The thesis also investigates different options for ac- cessing the hardware video encoder on Google Glass. Sammanfattning Internet har dramatiskt ändrat hur vi kommunicerar, och det blir allt viktigare för kommunikationssystem att kunna anpassa sig till kontexten som de används i. Målet med det här examensarbetet var att undersöka hur Google Glass och WebRTC kan användas för att skapa ett kommunikationssystem som är skräddarsytt för support av fälttekniker. -
News Release for More Information: Neil Trevett, President, Khronos | [email protected] | Phone: +1 (408) 464 7053
News Release For more information: Neil Trevett, President, Khronos | [email protected] | Phone: +1 (408) 464 7053 Khronos Releases OpenVX 1.1 Specification for High Performance, Low Power Computer Vision Acceleration Expanded range of processing functions; Enhanced flexibility for data access and processing; Full conformance tests available; Safety Critical specification in development May 2nd 2016 – Embedded Vision Summit, Santa Clara, CA – The Khronos™ Group, an open consortium of leading hardware and software companies, announces the immediate availability of the OpenVX™ 1.1 specification for cross platform acceleration of computer vision applications and libraries. OpenVX enables performance and power optimized computer vision algorithms for use cases such as face, body and gesture tracking, smart video surveillance, automatic driver assistance systems, object and scene reconstruction, augmented reality, visual inspection, robotics and more. Conformant OpenVX 1.0 implementations and tools are shipping from AMD, Imagination, Intel, NVIDIA, Synopsis and VeriSilicon. OpenVX 1.1 builds on this momentum by adding new processing functions for use cases such as computational photography, and enhances application control over how data is accessed and processed. An open source OpenVX 1.1 sample implementation and full conformance tests will be available before mid-2016. Details on the OpenVX specifications and Adopters Program are available at: www.khronos.org/openvx. “More and more products are incorporating computer vision, and OpenVX -
Tms320c6a8168, Tms320c6a8167 C6000 Dsp+Arm
TMS320C6A8168 TMS320C6A8167 www.ti.com SPRS680C –OCTOBER 2010–REVISED MARCH 2012 TMS320C6A816x C6000 DSP+ARM Processors Check for Samples: TMS320C6A8168, TMS320C6A8167 1 Device Summary 1.1 Features 1234567891011 • High-Performance C6000™ DSP+ARM® Multiply Supported up to: Processors – 2 SP x SP → SP Per Clock – ARM® Cortex™-A8 RISC Processor – 2 SP x SP → DP Every Two Clocks • Up to 1.5 GHz – 2 SP x DP → DP Every Three Clocks – C674x VLIW DSP – 2 DP x DP → DP Every Four Clocks • Up to 1.25 GHz • Fixed-Point Multiply Supports Two 32 x • Up to 12000/9000 C674x MIPS/MFLOPS 32 Multiplies, Four 16 x 16-bit Multiplies • Fully Software-Compatible with C67x+™ including Complex Multiplies, or Eight 8 x and C64x+™ 8-Bit Multiplies per Clock Cycle • ARM® Cortex™-A8 Core • C674x Two-Level Memory Architecture – ARMv7 Architecture – 32K-Byte L1P and L1D RAM/Cache • In-Order, Dual-Issue, Superscalar – 256K-Byte L2 Unified Mapped RAM/Caches Processor Core • DSP/EDMA Memory Management Unit • NEON™ Multimedia Architecture (DEMMU) – Supports Integer and Floating Point (VFPv3- – Maps C674x DSP and EMDA TCB Memory IEEE754 compliant) Accesses to System Addresses • Jazelle® RCT Execution Environment • 512K-Bytes On-Chip Memory Controller (OCMC) RAM • ARM® Cortex™-A8 Memory Architecture • SGX530 3D Graphics Engine (available only on – 32K-Byte Instruction and Data Caches the C6A8168 device) – 256K-Byte L2 Cache – Delivers up to 30 MTriangles/s – 64K-Byte RAM, 48K-Byte Boot ROM – Universal Scalable Shader Engine • TMS320C674x Floating-Point VLIW DSP – Direct3D®