Motionhub: Middleware for Unification of Multiple Body Tracking Systems

Total Page:16

File Type:pdf, Size:1020Kb

Motionhub: Middleware for Unification of Multiple Body Tracking Systems MotionHub: Middleware for Unification of Multiple Body Tracking Systems Philipp Ladwig Kester Evers Eric J. Jansen Mixed Reality and Visualization Mixed Reality and Visualization Mixed Reality and Visualization Group (MIREVI) Group (MIREVI) Group (MIREVI) University of Applied Sciences University of Applied Sciences University of Applied Sciences Düsseldorf, Germany Düsseldorf, Germany Düsseldorf, Germany [email protected] [email protected] [email protected] Ben Fischer David Nowottnik Christian Geiger Mixed Reality and Visualization Mixed Reality and Visualization Mixed Reality and Visualization Group (MIREVI) Group (MIREVI) Group (MIREVI) University of Applied Sciences University of Applied Sciences University of Applied Sciences Düsseldorf, Germany Düsseldorf, Germany Düsseldorf, Germany [email protected] [email protected] [email protected] Figure 1: MotionHub is an open-source middleware that offers interfaces to multiple body tracking systems. a) Two users were captured and tracked by a Microsoft Azure Kinect. b) The graphical user interface of MotionHub. The green cubes represent an OptiTrack recording while the yellow ones represent an Azure Kinect live capture. c) MotionHub streams unified skeletal representation in real time to clients such as the Unity game engine via a plug-in. ABSTRACT to another system can be a complex procedure. In this paper, we There is a substantial number of body tracking systems (BTS), which present our middleware solution MotionHub, which can receive and cover a wide variety of different technology, quality and price range process data of different BTS technologies. It converts the spatial for character animation, dancing or gaming. To the disadvantage as well as the skeletal tracking data into a standardized format in of developers and artists, almost every BTS streams out different real time and streams it to a client (e.g. a game engine). That way, protocols and tracking data. Not only do they vary in terms of scale MotionHub ensures that a client always receives the same skeletal- and offset, but also their skeletal data differs in rotational offsets data structure, irrespective of the used BTS. As a simple interface between joints and in the overall number of bones. Due to this enabling the user to easily change, set up, calibrate, operate and circumstance, BTSs are not effortlessly interchangeable. Usually, benchmark different tracking systems, the software targets artists software that makes use of a BTS is rigidly bound to it, and a change and technicians. MotionHub is open source, and other developers are welcome to contribute to this project. Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation CCS CONCEPTS on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or • Computing methodologies → Motion processing; • Infor- republish, to post on servers or to redistribute to lists, requires prior specific permission mation systems → Multimedia streaming; and/or a fee. Request permissions from [email protected]. MOCO ’20, July 15–17, 2020, Jersey City/ Virtual, NJ, USA © 2020 Copyright held by the owner/author(s). Publication rights licensed to the KEYWORDS Association for Computing Machinery. Body tracking, middleware, skeletal data, motion capture, Azure ACM ISBN 978-1-4503-7505-4/20/07...$15.00 https://doi.org/10.1145/3401956.3404185 Kinect, OptiTrack MOCO ’20, July 15–17, 2020, Jersey City/ Virtual, NJ, USA Ladwig et al. ACM Reference Format: A demo video with a summary of MotionHub’s capabilities can Philipp Ladwig, Kester Evers, Eric J. Jansen, Ben Fischer, David Nowottnik, be watched here: https://youtu.be/7caURXod-ag. and Christian Geiger. 2020. MotionHub: Middleware for Unification of Mul- The Git repository can be found here: https://github.com/Mirevi/ tiple Body Tracking Systems. In 7th International Conference on Movement MotionHub. and Computing (MOCO ’20), July 15–17, 2020, Jersey City/ Virtual, NJ, USA. ACM, New York, NY, USA, 8 pages. https://doi.org/10.1145/3401956.3404185 2 RELATED WORK 1 INTRODUCTION Body tracking is a wide field with many years of research and development, and it has produced a large number of different soft- Real-time body tracking is an indispensable part of many interac- ware and hardware approaches as well as standardizations and file tive art performances, mixed-reality applications and games. Due formats. In this section, we only mention the most common and to the progress in research, higher computing power and advances important ones for MotionHub. in sensor technology, a large number of BTSs that are suitable for such applications have been developed in recent years. Develop- 2.1 Standards and File Formats ers are spoilt for choice: They have to select a system and must consider various advantages and disadvantages concerning price, The probably best known and oldest de facto standard for data accuracy and size of the tracking area. However, in many cases exchange in virtual reality is the Virtual-Reality Peripheral Network not all requirements are known at the beginning of a project and (VRPN) by Taylor et al. [42]. It offers simple and unified interfaces will only be developed over time. This can pose a challenge as a to a broad range of devices of different manufacturers. Many of subsequent change to another tracking system can be costly. In these devices share common functionalities such as 6-DOF tracking such a scenario, a middleware that allows for an effortless switch or button input while the way of accessing these functions differs between different BTSs would be useful. between manufacturers. VRPN unifies functions across different The term ’middleware’ was probably first used in 1968 [35]. Since devices as generic classes such as vrpn_Tracker or vrpn_Button. then, various sub types have been described, but the basic meaning Therefore, it can be seen as both a standard and a middleware. The has never changed: A middleware receives, processes and transmits approach of MotionHub is similar, but it focuses on body tracking. information between at least two peers. Middlewares are often able The first official international standard for humanoid animation to understand multiple information representations and to translate is the H-Anim, which was created within the scope of the Extensible them into a standardized or unified one. 3D (X3D) standard and is a successor of the Virtual Reality Model- The contribution of this paper is the open-source middleware ing Language (VRML) [13, 14]. H-Anim was published in 2006 [15], ’MotionHub’ and an open-source game engine plugin, which can updated in 2019 [16, 17] and is one of the only efforts yet to create be seen in Fig. 1b and c. The intention behind our work is to unify an official open standard for humanoid avatar motion and data the output data of the wide variety of available BTSs. We believe exchange. that a unification will not only lead to an easier switch between COLLADA and FBX are interchange file formats for 3D appli- BTSs during production but that it will also reduce the preparation cations and are widely used today. While humanoid animation is time for developments that involve BTSs. Integrating a BTS into an not the focus of COLLADA, its open and versatile structure enables existing application requires a manual effort. Then, if developers developers to save body tracking data. Compared to COLLADA, the want to integrate another BTS or exchange the previous one with proprietary FBX format mainly focuses on motion data but lacks a it, an additional effort must be considered. In some cases, an update clear documentation, which has led to incompatible versions. to a newer version of a BTS is desired, which produces an even While COLLADA and FBX can also be used for writing and higher manual effort. Our intention is to integrate a generic BTS reading 3D geometry, the Biovision Hierarchy (BVH) file format protocol into applications once in order to reduce the overall time was developed exclusively for handling skeletal motion data and spent for maintenance, the setup and the switch between different is therefore simpler in structure. It is supported by many body BTSs. tracking applications and, because of its simplicity and less over- Beyond this, more benefits can be mentioned. If a middleware head compared to other file formats, it is often used for real-time understands different BTSs, it is possible to prioritize and to switch transmission of humanoid motion data. A deeper and more com- between them automatically to exploit the advantages of each sys- prehensive overview of further tracking file formats is given by tem. A possible scenario could be the use of a BTS that is accurate Meredith and Maddock [30]. but limited in tracking space and a different one that has a larger tracking space but is less accurate. A middleware could prioritize 2.2 Software and Hardware the more accurate BTS as long as the tracked person stays in the Microsoft started shipping the Kinect in 2010 and thereby allowed ’accurate’ tracking space and automatically switches to the ’less the body tracking community to grow substantially since the Kinect accurate’ space whenever he or she leaves it. was an affordable and capable sensor. During this time, PrimeSense Despite the advantages, the concept of the MotionHub also has a made OpenNI [39] and NITE [39, p.15] publicly available. OpenNI of- major drawback: By nature, a middleware induces a delay because fers low-level access to the Microsoft Kinect and other PrimeSense it receives, converts and resends the data between two ends, which sensors, while NITE was a middleware that enables the user to requires a certain processing time.
Recommended publications
  • The Application of Virtual Reality in Engineering Education
    applied sciences Review The Application of Virtual Reality in Engineering Education Maged Soliman 1 , Apostolos Pesyridis 2,3, Damon Dalaymani-Zad 1,*, Mohammed Gronfula 2 and Miltiadis Kourmpetis 2 1 College of Engineering, Design and Physical Sciences, Brunel University London, London UB3 3PH, UK; [email protected] 2 College of Engineering, Alasala University, King Fahad Bin Abdulaziz Rd., Dammam 31483, Saudi Arabia; [email protected] (A.P.); [email protected] (M.G.); [email protected] (M.K.) 3 Metapower Limited, Northwood, London HA6 2NP, UK * Correspondence: [email protected] Abstract: The advancement of VR technology through the increase in its processing power and decrease in its cost and form factor induced the research and market interest away from the gaming industry and towards education and training. In this paper, we argue and present evidence from vast research that VR is an excellent tool in engineering education. Through our review, we deduced that VR has positive cognitive and pedagogical benefits in engineering education, which ultimately improves the students’ understanding of the subjects, performance and grades, and education experience. In addition, the benefits extend to the university/institution in terms of reduced liability, infrastructure, and cost through the use of VR as a replacement to physical laboratories. There are added benefits of equal educational experience for the students with special needs as well as distance learning students who have no access to physical labs. Furthermore, recent reviews identified that VR applications for education currently lack learning theories and objectives integration in their design.
    [Show full text]
  • BOOK of NEWS Al in Business
    BOOK OF NEWS Microsoft Inspire July 14 – 18, 2019 Las Vegas, NV Contents Foreword by Gavriella Schuster 4 Inspire Overview and How to Watch 5 Chapter 1 6 Customer and Partner Enablement Item 1.1 Partner program and investments update Item 1.2 Microsoft Azure Lighthouse Item 1.3 Microsoft Edge Enterprise Readiness Item 1.4 Microsoft Intelligent Security Association expands Chapter 2 9 Microsoft Business Applications | ISV Opportunity Item 2.1 New Microsoft Business Applications ISV Connect program and technology industry innovation for ISVs Item 2.2 Industry accelerators for ISVs Item 2.3 Latest version of the Dynamics 365 Nonprofit Accelerator Chapter 3 13 Marketplace Item 3.1 New monetization models, rewards program and route to market in commercial marketplace Chapter 4 15 Microsoft Azure and AI Item 4.1 AI for Cultural Heritage Item 4.2 Cloud Adoption Framework Item 4.3 AI Accelerate Program Item 4.4 New Azure migration capabilities Item 4.5 Azure Migration Program Item 4.6 Azure Data Box Heavy is now generally available; Data Box and Data Box Disk regions are expanded Item 4.7 Open source Quantum Development Kit Item 4.8 Azure Kinect DK General Availability Item 4.9 ServiceNow makes Azure Cloud its preferred cloud platform for highly regulated industries Contents | 2 Contents Chapter 5 20 Databases and Analytics Item 5.1 Blob API interoperability with Azure Data Lake Storage Gen2 Public Preview Item 5.2 Azure Data Share, new analytics service, announced for enterprises Chapter 6 22 Microsoft Teams Item 6.1 Microsoft Teams reaches 13
    [Show full text]
  • Design Architecture in Virtual Reality
    Design Architecture in Virtual Reality by Anisha Sankar A thesis presented to the University of Waterloo in fulfillment of the thesis requirement for the degree of Master of Architecture Waterloo, Ontario, Canada, 2019 © Anisha Sankar 2019 Author’s Declaration I hereby declare that I am the sole author of this thesis. This is a true copy of the thesis, including any required final revisions, as accepted by my examiners. I understand that my thesis may be made electronically available to the public. - iii - Abstract Architectural representation has newly been introduced to Virtual Real- ity (VR) technology, which provides architects with a medium to show- case unbuilt designs as immersive experiences. Designers can use specialized VR headsets and equipment to provide a client or member of their design team with the illusion of being within the digital space they are presented on screen. This mode of representation is unprec- edented to the architectural field, as VR is able to create the sensation of being encompassed in an environment at full scale, potentially elic- iting a visceral response from users, similar to the response physical architecture produces. While this premise makes the technology highly applicable towards the architectural practice, it might not be the most practical medium to communicate design intent. Since VR’s conception, the primary software to facilitate VR content creation has been geared towards programmers rather than architects. The practicality of inte- grating virtual reality within a traditional architectural design workflow is often overlooked in the discussion surrounding the use of VR to rep- resent design projects. This thesis aims to investigate the practicality of VR as part of a de- sign methodology, through the assessment of efficacy and efficiency, while studying the integration of VR into the architectural workflow.
    [Show full text]
  • Learn More: Windows Mixed Reality Platform + Steamvr
    NOT FOR USE IN INDIA FAQ HP REVERB VR HEADSET PROFESSIONAL EDITION, STEAMVR & WINDOWS MIXED REALITY 2 1.0 STEAMVR 3 2.0 HP REVERB VR HEADSET PRO EDITION 6 3.0 WINDOWS MIXED REALITY 7 LEARN MORE: WINDOWS MIXED 4.0 GENERAL VR REALITY PLATFORM + STEAMVR Frequently asked questions about HP’s professional head-mounted display (HMD) - built on the Windows Mixed Reality (WMR) platform - and integration with SteamVR. The HP Reverb Virtual Reality Headset - Professional Edition offers stunning immersive computing with significant ease of setup and use in a cost effective solution. This solution is well suited for Engineering Product Dev and design reviews, AEC (Architecture, Engineering & Construction) reviews, location-based entertainment, and MRO (Maintenance, Repair and Overhaul) training use environments. HIGHLIGHT: Take advantage of the complete Windows 10 Mixed Reality and SteamVR ecosystems. The HP Reverb VR Headset Pro HP Reverb Virtual Reality Headset - Professional Edition is not recommended for children under the age of Edition is built on the Windows IMPORTANT NOTE: 13. All users should read the HP Reverb Virtual Reality Headset - Professional Edition User Guide to reduce the risk of personal Mixed Reality platform. injury, discomfort, property damage, and other potential hazards and for important information related to your health and Integration with SteamVR safety when using the headset. Windows Mixed Reality requires Windows 10 October 2018 Update installed on the workstation requires the Windows Mixed or PC. Features may require software or other 3rd-party applications to provide the described functionality. To minimize the Reality bridge app. possibility of experiencing discomfort using a VR application, ensure that the PC system is equipped with the appropriate graphics and CPU for the VR application.
    [Show full text]
  • Microsoft 2012 Citizenship Report
    Citizenship at Microsoft Our Company Serving Communities Working Responsibly About this Report Microsoft 2012 Citizenship Report Microsoft 2012 Citizenship Report 01 Contents Citizenship at Microsoft Serving Communities Working Responsibly About this Report 3 Serving communities 14 Creating opportunities for youth 46 Our people 85 Reporting year 4 Working responsibly 15 Empowering youth through 47 Compensation and benefits 85 Scope 4 Citizenship governance education and technology 48 Diversity and inclusion 85 Additional reporting 5 Setting priorities and 16 Inspiring young imaginations 50 Training and development 85 Feedback stakeholder engagement 18 Realizing potential with new skills 51 Health and safety 86 United Nations Global Compact 5 External frameworks 20 Supporting youth-focused 53 Environment 6 FY12 highlights and achievements nonprofits 54 Impact of our operations 23 Empowering nonprofits 58 Technology for the environment 24 Donating software to nonprofits Our Company worldwide 61 Human rights 26 Providing hardware to more people 62 Affirming our commitment 28 Sharing knowledge to build capacity 64 Privacy and data security 8 Our business 28 Solutions in action 65 Online safety 8 Where we are 67 Freedom of expression 8 Engaging our customers 31 Employee giving and partners 32 Helping employees make 69 Responsible sourcing 10 Our products a difference 71 Hardware production 11 Investing in innovation 73 Conflict minerals 36 Humanitarian response 74 Expanding our efforts 37 Providing assistance in times of need 76 Governance 40 Accessibility 77 Corporate governance 41 Empowering people with disabilities 79 Maintaining strong practices and performance 42 Engaging students with special needs 80 Public policy engagement 44 Improving seniors’ well-being 83 Compliance Cover: Participants at the 2012 Imagine Cup, Sydney, Australia.
    [Show full text]
  • Immersive Virtual Reality Attacks and the Human Joystick Peter Casey University of New Haven
    University of New Haven Masthead Logo Digital Commons @ New Haven Electrical & Computer Engineering and Computer Electrical & Computer Engineering and Computer Science Faculty Publications Science 3-27-2019 Immersive Virtual Reality Attacks and the Human Joystick Peter Casey University of New Haven Ibrahim Baggili University of New Haven, [email protected] Ananya Yarramreddy University of New Haven Follow this and additional works at: https://digitalcommons.newhaven.edu/ electricalcomputerengineering-facpubs Part of the Computer Engineering Commons, Electrical and Computer Engineering Commons, Forensic Science and Technology Commons, and the Information Security Commons Publisher Citation Casey, P., Baggili, I., & Yarramreddy, A. (2019). Immersive Virtual Reality Attacks and the Human Joystick. IEEE Transactions on Dependable and Secure Computing. doi:10.1109/TDSC.2019.2907942 Comments This material is based upon work supported by the National Science Foundation under Grant No. 1748950. © © 2019 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works. The ev rsion of record may be found at http://dx.doi.org/10.1109/ TDSC.2019.2907942. Dr. Baggili was appointed to the University of New Haven's Elder Family Endowed Chair in 2015. 1 Immersive Virtual Reality Attacks and the Human Joystick Peter Casey, Ibrahim Baggili, and Ananya Yarramreddy Abstract—This is one of the first accounts for the security analysis of consumer immersive Virtual Reality (VR) systems.
    [Show full text]
  • Azure-Kinect and Augmented Reality for Learning Basic Mathematics - a Case Study
    Preprints (www.preprints.org) | NOT PEER-REVIEWED | Posted: 30 September 2020 doi:10.20944/preprints202009.0752.v1 Azure-Kinect and Augmented Reality for learning Basic Mathematics - A case study Raúl Lozada-Yánez 1, 2, Nora La-Serna Palomino 1, Diego Veloz-Chérrez 2, Fernando Molina-Granja 3 & Juan Carlos Santillán-Lima 4 1 Universidad Nacional Mayor de San Marcos, Facultad de Ingeniería de Sistemas e Informática, [email protected], [email protected] 2 Escuela Superior Politécnica de Chimborazo, [email protected],Riobamba, Ecuador Facultad de Informática en Electrónica, [email protected], [email protected] 3 Universidad Nacional de Chimborazo, Riobamba, Ecuador Facultad de Ingeniería,[email protected] 4 Universidad Nacional de La Plata, Ciudad de la Plata, Argentina Facultad de Informática - LEICI - Instituto de Investigaciones en Electrónica, Control y Procesamiento de Se- ñales. Facultad de Ingeniería, [email protected], [email protected] *Correspondence: [email protected]; [email protected] Abstract: The way in which the human being learns certain complex contents has always been a focus of interest and a challenge for researchers. Given the fact that children's cognitive abilities do not fully develop until a certain age, this topic is particularly important in the young children's learning scope as they do not correctly and easily learn some content of abstract nature, such as contents in math class. This work presents the results of the use of an application called "Mathematics Learning System with Augmented Reality based on Kinect" (SAM-RAK by its acronym in Spanish), which was designed to cover basic topics of mathematics in the Basic General Education level (EGB by its acronym in Spanish) in Ecuador.
    [Show full text]
  • Azure Kinect Examples for Unity
    Azure Kinect Examples for Unity Copyright © RF Solutions. All rights reserved. Azure Kinect Examples for Unity Table of contents Introduction ........................................................................................................................ 3 How to Run ........................................................................................................................ 3 Azure Kinect SDKs ............................................................................................................. 3 Kinect-v2 SDK ................................................................................................................... 3 RealSense SDK .................................................................................................................. 4 Demo Scenes ...................................................................................................................... 4 VFX Point-Cloud Demo ......................................................................................................... 8 The Kinect Manager ............................................................................................................. 9 Sensor Interfaces .............................................................................................................. 10 General-Purpose Components ............................................................................................. 10 Demo-Specific Components ................................................................................................
    [Show full text]
  • Self Representation and Interaction in Immersive Virtual Reality
    Self Representation and Interaction in Immersive Virtual Reality Eros Viola a, Fabio Solari b and Manuela Chessa c Dept. of Informatics, Bioengineering, Robotics, and Systems Engineering, University of Genoa, Italy Keywords: Virtual Reality, Self Representation, Intel Realsense D435, Interaction, Leap Motion, Manus Prime Haptic Gloves, Alignment, Rigid Transformation, Live Correction. Abstract: Inserting a self-representation in Virtual Reality is an open problem with several implications for both the sense of presence and interaction in virtual environments. To cope the problem with low cost devices, we de- vise a framework to align the measurements of different acquisition devices used while wearing a tracked VR head-mounted display (HMD). Specifically, we use the skeletal tracking features of an RGB-D sensor (Intel Realsense d435) to build the user’s avatar, and compare different interaction technologies: a Leap Motion, the Manus Prime haptic gloves, and the Oculus Controllers. The effectiveness of the proposed systems is assessed through an experimental session, where an assembly task is proposed with the three different interaction medi- ums, with and without the self-representation. Users reported their feeling by answering the User Experience and Igroup Presence Questionnaires, and we analyze the total time to completion and the error rate. 1 INTRODUCTION The focus is to develop a framework that is compat- ible with the most common head-mounted displays In current Virtual Reality (VR) applications, the vi- (HMDs), and that can also be extended to other track- sual feedback of the user’s body is often missing, de- ing devices, both for the body tracking and for the spite the abundance of enabling technologies, which hands and fingers tracking.
    [Show full text]
  • Subject Index
    863 Subject Index ‘Note: Page numbers followed by “f” indicate figures, “t” indicate tables and “b” indicate boxes.’ A Affordances, 112–114 A/D conversion. See Analog to Digital conversion in virtual reality, 114–117 (A/D conversion) false affordances, 116 AAAD. See Action at a distance (AAAD) reinforcing perceived affordances, 116–117 AAR. See After-action review (AAR) After-action review (AAR), 545f, 630, 630f, 634–635, 645, Absolute input, 198–200 761 Abstract haptic representations, 439 Affordances of VR, 114–117 Abstract synthesis, 495 Agency, 162, 164, 181 Abstraction triangle, 448 Agents, 552–553, 592–593, 614, 684–685 Accelerometers, 198–199, 218 AIFF. See Audio interchange file format (AIFF) Accommodation, 140–141, 273–275, 320, 570, 804 Airfoils, 13 Action at a distance (AAAD), 557–558 AIs. See Artificial intelligences (AIs) Activation mechanism, 554–556, 583 Aladdin’s Magic Carpet Ride VR experience, 185–186, 347, Active haptic displays, 516 347f, 470, 505, 625, 735, 770–771 Active input, 193–196 Alberti, Leon Battista, 28 Active surfaces, 808 Alice system for programming education, Adaptability, 122–123 758–759 Adaptive rectangular decomposition (ARD), 502 AlloSphere, 51–52, 51f, 280 Additive sound creation techniques, 499 Allstate Impaired Driver Simulator, 625, 629f Advanced Realtime Tracking (ART), 53–54, 213f Alpha delta fiber (Aδ fiber), 149 Advanced Robotics Research Lab (ARRL), 369 Alphanumeric value selection, 591–593 Advanced systems, 11–12 Ambient sounds, 436–437, 505 Advanced texture mapping techniques, 469–473 Ambiotherm device, 372, 373f Adventure (games), 12 Ambisonics, 354 Adverse effect, 351 Ambulatory platforms, 242–243 Aestheticism, 414 American Sign Language (ASL), 552–553 Affine transformations, 486–487 Amount/type of information, 196–198 864 | SUBJECT INDEX Amplification, 349–350 ARToolKit (ARTK), 48–49, 715–716 Amplifier, 349–350, 349f Ascension Technologies, 41–44, 44f, 46–47, 86–87 Anaglyphic 3D, 270f, 7f, 30, 49, 269–270, 271f ASL.
    [Show full text]
  • Envrment: Investigating Experience in a Virtual User-Composed Environment
    ENVRMENT: INVESTIGATING EXPERIENCE IN A VIRTUAL USER-COMPOSED ENVIRONMENT A Thesis presented to the Faculty of California Polytechnic State University, San Luis Obispo In Partial Fulfillment of the Requirements for the Degree Master of Science in Computer Science by Matthew Key December 2020 c 2020 Matthew Key ALL RIGHTS RESERVED ii COMMITTEE MEMBERSHIP TITLE: EnVRMent: Investigating Experience in a Virtual User-Composed Environment AUTHOR: Matthew Key DATE SUBMITTED: December 2020 COMMITTEE CHAIR: Zo¨eWood, Ph.D. Professor of Computer Science COMMITTEE MEMBER: Christian Eckhardt, Ph.D. Professor of Computer Science COMMITTEE MEMBER: Franz Kurfess, Ph.D. Professor of Computer Science iii ABSTRACT EnVRMent: Investigating Experience in a Virtual User-Composed Environment Matthew Key Virtual Reality is a technology that has long held society's interest, but has only recently began to reach a critical mass of everyday consumers. The idea of modern VR can be traced back decades, but because of the limitations of the technology (both hardware and software), we are only now exploring its potential. At present, VR can be used for tele-surgery, PTSD therapy, social training, professional meetings, conferences, and much more. It is no longer just an expensive gimmick to go on a momentary field trip; it is a tool, and as with the automobile, personal computer, and smartphone, it will only evolve as more and more adopt and utilize it in various ways. It can provide a three dimensional interface where only two dimensions were previously possible. It can allow us to express ourselves to one another in new ways regardless of the distance between individuals.
    [Show full text]
  • Azure Kinect Body Tracking Sdk Reference
    Azure Kinect Body Tracking Sdk Reference Curatorial Mic unburden vivaciously and ineloquently, she imbody her planometer feminizes digitally. Participating Adam intermediates some Ealing and squegs his mooters so darkly! Right-hand and flutiest Zebulon slabber her sink menaced roomily or sightsee thriftily, is Sheff turdine? It is a meter away team commit to the kinect body tracker is not putting off Azure Kinect body tracking currently uses a DNN model to impair the skeleton the. Azure Kinect Masterclass Angles LightBuzz. Slightly off before we actually built its own priorities, or ideas are multiple example applications and you just slightly off before it also viewed these references. Net framework development, but i wanted with. My money when depth. Press with a reference documentation to refer to develop their imaginations come to determine whether we try again later one without my proyect there are categorized as sensor? Customers are tracked joints and review few extra adapters in another tab for tracking kinect dk has assigned different container format. MicrosoftAzureKinectBodyTracking 101 NuGet Gallery. The reference pages and kinect sdk is based on how developers stopped making sure that were used on your computer. Eddie using azure kinect or. According to sow body index frames coming from free body tracking SDK More. See details on hardware recommendations in Microsoft's Azure Kinect documentation. For more information visit the official documentation of Azure Kinect DK. Add C wrapper for Azure Kinect Body Tracking SDK Azure. Kinect camera to play them with so. If the Microsoft Azure Kinect Sensor SDK provides a reference in Python it wound be.
    [Show full text]