Haptic Interactions in Virtual Reality MQP

Total Page:16

File Type:pdf, Size:1020Kb

Haptic Interactions in Virtual Reality MQP Haptic Interactions in Virtual Reality MQP Major Qualifying Project Report: Submitted to the Faculty of In partial fulfillment of the requirements for the Degree of Bachelor of Science In association with By _________________ Richard Matthew Rafferty RBE _________________ Jiaqi Ren RBE/ME _________________________ __________________________ Professor Gregory S. Fischer Advisor Professor Robert W. Lindeman Co-Advisor Page | 1 Table of Contents Abstract ................................................................................................................................................. 5 Executive Summary ............................................................................................................................... 6 I. Background ................................................................................................................................. 6 II. Functional Requirements ............................................................................................................. 6 III. System Design ......................................................................................................................... 7 IV. Results ..................................................................................................................................... 7 V. Discussion ................................................................................................................................... 8 1. Introduction ................................................................................................................................. 9 2. Background Research ................................................................................................................ 10 2.1. Virtual Reality ....................................................................................................................... 10 2.2. Haptics .................................................................................................................................. 14 2.3. Rehabilitation ........................................................................................................................ 16 2.4. Human Hand Dynamics ......................................................................................................... 18 2.5. Current Solutions ................................................................................................................... 18 2.5.1. CyberGrasp and Virtual Hand SDK by CyberGlove Systems{CyberGlove, 2015 #10} ... 19 2.5.2. Falcon by Novint {Novint, 2015 #12} ............................................................................ 20 2.5.3. Geomagic Haptic Devices {Geomagic, #25} .................................................................. 20 3. Methodology ............................................................................................................................. 22 3.1. Project definition ................................................................................................................... 22 3.2. System Requirements ............................................................................................................ 24 3.3. Sub-Systems Design Decisions .............................................................................................. 25 3.3.1. Hand Pose Tracking System ........................................................................................... 25 3.3.2. Simulation System .......................................................................................................... 26 3.3.3. Force Exertion System: Mechanical Structure ................................................................. 27 Page | 2 3.3.4. Control System ............................................................................................................... 30 3.3.5. Communication Design Decision .................................................................................... 32 3.3.6. Virtual Environment Viewing System ............................................................................. 35 3.3.7. Simulation System and Force Exertion System Communication ..................................... 36 3.4. Demo Application: Rehabilitating Hand Learned Non-Use through Visual Feedback Distortion 38 3.4.1. Procedure Steps .............................................................................................................. 38 3.4.2. Demo Requirements ....................................................................................................... 39 4. Final Design .............................................................................................................................. 40 4.1. System Implementation ......................................................................................................... 40 4.2. Force Exertion System ........................................................................................................... 40 4.2.1. Bowden Cable Exoskeleton ............................................................................................ 42 4.2.2. DC Brushed Motors ........................................................................................................ 43 4.2.3. Microcontroller ............................................................................................................... 44 4.2.4. Motor Controller Shield .................................................................................................. 45 4.2.5. Current Sensing Module ................................................................................................. 46 4.2.6. Force Exertion Control ................................................................................................... 48 4.3. Hand Pose Tracking System .................................................................................................. 49 4.4. Simulation System ................................................................................................................. 50 4.4.1. Environment and Development Tools ............................................................................. 51 4.4.2. User and Object Representations..................................................................................... 51 4.4.3. Interaction Forces ........................................................................................................... 52 4.4.4. Force Communication ..................................................................................................... 54 4.5. Visualization System ............................................................................................................. 57 4.6. Demonstrative Application: Learned Non-Use Rehabilitation ................................................ 58 Page | 3 .............................................................................................................................Results and Discussion ............................................................................................................................................................ 59 5. Results and Discussion ..................................................................................................................... 59 5.1. Hand Pose Tracking System Functional Analysis ................................................................... 59 5.2. Force Exertion System Functional Analysis ........................................................................... 60 5.2.1. Exoskeleton Stiffness ...................................................................................................... 63 5.2.2. Exoskeleton Safety ......................................................................................................... 64 5.3. Simulation System Functional Analysis ................................................................................. 64 5.3.1. Haptic Interaction Generation ......................................................................................... 65 5.4. Serial Communication Issue and Sub-System Integration ....................................................... 65 5.5. Visualization System Functional Analysis ............................................................................. 67 5.6. Demo Application Functional Analysis .................................................................................. 68 5.7. Development Guide ............................................................................................................... 68 5.8. Overall System Functional Analysis ...................................................................................... 69 6. Future Work .............................................................................................................................. 70 6.1. Sub-System Enhancement ...................................................................................................... 70 6.2. Virtual Reality Presence Sub-Systems ................................................................................... 70 6.3. Application Development ...................................................................................................... 70 6.4. System Marketing and Distribution ........................................................................................ 71 7. Conclusion ................................................................................................................................ 72 8. Bibliography ............................................................................................................................
Recommended publications
  • Virtual and Augmented Reality in Marketing
    Virtual and Augmented Reality in Marketing Laura Håkansson Thesis Tietojenkäsittelyn Koulutusohjelma 2018 Authors Group Laura Håkansson HETI15KIM The title of your thesis Number of Virtual and Augmented Reality in Marketing pages and appendices 56 Supervisors Heikki Hietala This thesis serves as an introduction to Virtual and Augmented Reality and explains how these two different technologies can be used in marketing. I work in marketing and have been following the evolvement of both VR and AR for a few years now. I was personally very curious to learn about how these have been implemented in marketing in the past, and what they will look like in the future, and if there are any reoccurring themes, for both VR and AR, that work best with a specific product or service. I spent a year reading the latest news and articles about various VR/AR marketing campaigns, and about the updates on these technologies that kept coming almost monthly. I also participated in different VR/AR-themed events in Helsinki to try out headsets and new AR apps, and to listen to the experts view on the potential of VR and AR. I wanted to create clear guidelines on which technology to use and how, depending on the product or service being marketed, but realized during my research that this was not possible. VR and AR are still in development, but evolving at a very fast pace, and right now it’s important to just bravely test them out without worrying about failing. I give plenty of examples in this thesis that will hopefully encourage marketers to start experimenting now, because we will see some really advanced VR and AR/MR in a few years, and those with the most experience will have great advantage in the marketing field.
    [Show full text]
  • 12.2% 116,000 125M Top 1% 154 4,200
    We are IntechOpen, the world’s leading publisher of Open Access books Built by scientists, for scientists 4,200 116,000 125M Open access books available International authors and editors Downloads Our authors are among the 154 TOP 1% 12.2% Countries delivered to most cited scientists Contributors from top 500 universities Selection of our books indexed in the Book Citation Index in Web of Science™ Core Collection (BKCI) Interested in publishing with us? Contact [email protected] Numbers displayed above are based on latest data collected. For more information visit www.intechopen.com Chapter 4 Waveguide-Type Head-Mounted Display System for AR Application Munkh-Uchral Erdenebat,Munkh-Uchral Erdenebat, Young-Tae Lim,Young-Tae Lim, Ki-Chul Kwon,Ki-Chul Kwon, Nyamsuren DarkhanbaatarNyamsuren Darkhanbaatar and Nam KimNam Kim Additional information is available at the end of the chapter http://dx.doi.org/10.5772/intechopen.75172 Abstract Currently, a lot of institutes and industries are working on the development of the virtual reality and augmented reality techniques, and these techniques have been recognized as the determination for the direction of the three-dimensional display development in the near future. In this chapter, we mainly discussed the design and application of several wearable head-mounted display (HMD) systems with the waveguide structure using the in- and out-couplers which are fabricated by the diffractive optical elements or holo - graphic volume gratings. Although the structure is simple, the waveguide-type HMDs are very efficient, especially in the practical applications, especially in the augmented reality applications, which make the device light-weighted.
    [Show full text]
  • Hate in Social VR New Challenges Ahead for the Next Generation of Social Media
    Hate in Social VR New Challenges Ahead for the Next Generation of Social Media Sections 1 Introduction: ADL and New Technologies 5 Cautionary Tales: Hate, Bias, and Harassment in Video Games, Social Media, and the Tech 2 Hopes and Fears for the Future of Social Virtual Industry Reality 6 Sexism, Racism, and Anti-Semitism in Social VR 3 Social VR: Industry Overview and Technological Today Affordances 7 What Comes Next for Social VR? 4 Hopeful Futures: Equity, Justice, and VR 8 Bibliography 9 Appendix: Codes of Conduct and Content Reporting and Moderation in Social VR 1 / 65 INTRODUCTION: ADL AND NEW TECHNOLOGIES The timeless mission of the Anti-Defamation League (ADL) is to stop the defamation of the Jewish people and secure justice and fair treatment to all. The mission of ADL’s Center for Technology and Society is to ask the question “How do we secure justice and fair treatment for all in a digital environment?” Since 1985, when it published its report on “Electronic Bulletin Boards of Hate,” ADL has been at the forefront of evaluating how new technologies can be used for good and misused to promote hate and harassment in society. At the same time, ADL has worked with technology companies to show them how they can address these issues and work to make the communities their products foster more respectful and inclusive. We continue this work today alongside Silicon Valley giants such as Facebook, Twitter, Google, Microsoft, Reddit, and others. This report represents a forward-looking review of the hopes and issues raised by the newest ecosystem of community fostering technology products: social virtual reality, or social VR.
    [Show full text]
  • Haptic Augmented Reality 10 Taxonomy, Research Status, and Challenges
    Haptic Augmented Reality 10 Taxonomy, Research Status, and Challenges Seokhee Jeon, Seungmoon Choi, and Matthias Harders CONTENTS 10.1 Introduction ..................................................................................................227 10.2 Taxonomies ...................................................................................................229 10.2.1 Visuo-Haptic Reality–Virtuality Continuum ...................................229 10.2.2 Artificial Recreation and Augmented Perception ............................. 232 10.2.3 Within- and Between-Property Augmentation ................................. 233 10.3 Components Required for Haptic AR ..........................................................234 10.3.1 Interface for Haptic AR ....................................................................234 10.3.2 Registration between Real and Virtual Stimuli ................................236 10.3.3 Rendering Algorithm for Augmentation ..........................................237 10.3.4 Models for Haptic AR.......................................................................238 10.4 Stiffness Modulation.....................................................................................239 10.4.1 Haptic AR Interface ..........................................................................240 10.4.2 Stiffness Modulation in Single-Contact Interaction .........................240 10.4.3 Stiffness Modulation in Two-Contact Squeezing .............................243 10.5 Application: Palpating Virtual Inclusion in Phantom
    [Show full text]
  • Teleoperation and Visualization Interfaces for In-Space Operations Will Pryor, Balazs P
    Teleoperation and Visualization Interfaces for In-Space Operations Will Pryor, Balazs P. Vagvolgyi, Anton Deguet, Simon Leonard, Louis L. Whitcomb, and Peter Kazanzides Johns Hopkins University, Baltimore, MD, USA Overview Teleoperation Interfaces • Ground-based teleoperation for on-orbit satellite servicing is complicated by communication delay and non-ideal camera views • We have evaluated teleoperation interfaces: • Direct teleoperation using da Vinci master console • Programmed motion primitives, where a single motion command is issued via a GUI • Model-mediated teleoperation, where the operator’s interaction with a simulated environment is replicated remotely • Interactive Planning and Supervised Execution (IPSE), where the user operates a virtual robot, evaluates the motion, and then supervises execution on the real robot • Keyboard, mouse and GUI • da Vinci Master Console: • Most similar to NASA interface • Stereo visualization • Visualization on multiple monitors • Two 3D input devices (da Vinci MTMs) • 3D monitor or HMD possible Direct teleoperation Motion primitives Model-mediated teleoperation Interactive Planning and Supervised Execution (IPSE) Experimental Tasks: Cutting multi-layer insulation (MLI) Mixed Reality Visualization Augmented Reality Augmented Virtuality Virtual Reality RRM Task Board • Virtual objects on real • Real camera images on • Virtual models OSAM-1 (formerly Restore-L) camera images virtual models Landsat7 MLI hat Scene Modeling and Registration • Manual construction Physical setup Cutting tool • Virtual
    [Show full text]
  • Project Final Report
    FP7-RSME-262160 PROJECT FINAL REPORT Publishable summary Grant Agreement number: 262160 Project acronym: Digital Ocean Project title: Integrated multimedia mixed reality system, of real time virtual diving, by web teleoperated underwater data collecting robots, diffused online and through a network of submersible simulation devices. Funding Scheme: Capacities - RSME Period covered: from 1st of January 2011 to 31st of December 2012. Name of the scientific representative of the project’s co-ordinator: Antonio Dinis Title and organisation: President, VirtualDive Tel: +33(0)1 34 62 83 89 E-mail: [email protected] Project website address: www.digitalocean.eu 1 FP7-RSME-262160 EXECUTIVE SUMMARY Introduction and rationale. 70% of our planet is covered by water and the very survival of more than 50% of humanity is conditioned by oceans. Nevertheless, only 0,4% of this population have ever seen, through actual diving, what exists under the sea, beneath water’s hiding surfaces. 25 million people are considered practicing as they dive, in average, a week per year. They come essentially from Europe and from the US, are generally middle class males averaging 35 years old and are motivated by adventurous aspects of diving. The number remains practically stable. Women, children, handicapped and seniors are barely accounted for in diving. Diving remains a risky, expensive and complicated activity for the majority of people, especially in emerging markets. In such conditions, the contribution of the diving community to the discovery of undersea or to global awareness of challenges facing the oceans remains limited. There is a sense of urgency about offering to the public, ways to discover by themselves, how oceans – earth’s fragile life support system, are being neglected, irreversibly polluted and impoverished.
    [Show full text]
  • Meaning–Making in Extended Reality Senso E Virtualità
    Meaning–Making in Extended Reality Senso e Virtualità edited by Federico Biggio Victoria Dos Santos Gianmarco Thierry Giuliana Contributes by Antonio Allegra, Federico Biggio, Mila Bujic,´ Valentino Catricalà Juan Chen, Kyle Davidson, Victoria Dos Santos, Ruggero Eugeni Gianmarco Thierry Giuliana, Mirko Lino, Alexiev Momchil Claudio Paolucci, Mattia Thibault, Oguz˘ ‘Oz’ Buruk Humberto Valdivieso, Ugo Volli, Nannan Xi, Zhe Xue Aracne editrice www.aracneeditrice.it [email protected] Copyright © MMXX Gioacchino Onorati editore S.r.l. – unipersonale www.gioacchinoonoratieditore.it [email protected] via Vittorio Veneto, Canterano (RM) () ---- No part of this book may be reproduced by print, photoprint, microfilm, microfiche, or any other means, without publisher’s authorization. Ist edition: June Table of Contents 9 Introduction Federico Biggio, Victoria Dos Santos, Gianmarco Thierry Giuliana Part I Making Sense of the Virtual 21 Archeologia semiotica del virtuale Ugo Volli 43 Una percezione macchinica: realtà virtuale e realtà aumentata tra simulacri e protesi dell’enunciazione Claudio Paolucci 63 Technologically Modified Self–Centred Worlds. Modes of Pres- ence as Effects of Sense in Virtual, Augmented, Mixed and Extend- ed Reality Valentino Catricalà, Ruggero Eugeni 91 Towards a Semiotics of Augmented Reality Federico Biggio 113 Deconstructing the Experience. Meaning–Making in Virtual Reality Between and Beyond Games and Films Gianmarco Thierry Giuliana, Alexiev Momchil 7 8 Table of Contents 143 The Digital and the Spiritual: Validating Religious Experience Through Virtual Reality Victoria Dos Santos 165 Role–play, culture, and identity in virtual space. Semiotics of digital interactions Kyle Davidson 189 Tecnica, virtualità e paura. Su una versione dell’angoscia contem- poranea Antonio Allegra Part II Extended Spaces and Realities 207 VRBAN.
    [Show full text]
  • Introduction to Virtual Reality
    Intelligent Systems, Control and Automation: Science and Engineering Volume 64 Series Editor S. G. Tzafestas For further volumes: http://www.springer.com/series/6259 Matjazˇ Mihelj • Janez Podobnik Haptics for Virtual Reality and Teleoperation 123 Matjazˇ Mihelj Janez Podobnik Faculty of Electrical Engineering Faculty of Electrical Engineering University of Ljubljana University of Ljubljana Ljubljana Ljubljana Slovenia Slovenia ISBN 978-94-007-5717-2 ISBN 978-94-007-5718-9 (eBook) DOI 10.1007/978-94-007-5718-9 Springer Dordrecht Heidelberg New York London Library of Congress Control Number: 2012951397 Ó Springer Science?Business Media Dordrecht 2012 This work is subject to copyright. All rights are reserved by the Publisher, whether the whole or part of the material is concerned, specifically the rights of translation, reprinting, reuse of illustrations, recitation, broadcasting, reproduction on microfilms or in any other physical way, and transmission or information storage and retrieval, electronic adaptation, computer software, or by similar or dissimilar methodology now known or hereafter developed. Exempted from this legal reservation are brief excerpts in connection with reviews or scholarly analysis or material supplied specifically for the purpose of being entered and executed on a computer system, for exclusive use by the purchaser of the work. Duplication of this publication or parts thereof is permitted only under the provisions of the Copyright Law of the Publisher’s location, in its current version, and permission for use must always be obtained from Springer. Permissions for use may be obtained through RightsLink at the Copyright Clearance Center. Violations are liable to prosecution under the respective Copyright Law.
    [Show full text]
  • Applications of Haptic Systems in Virtual Environments: a Brief Review
    Chapter 13 Applications of Haptic Systems in Virtual Environments: A Brief Review Alma G. Rodríguez Ramírez, Francesco J. García Luna, Osslan Osiris Vergara Villegas and Manuel Nandayapa Abstract Haptic systems and virtual environments represent two innovative technologies that have been attractive for the development of applications where the immersion of the user is the main concern. This chapter presents a brief review about applications of haptic systems in virtual environments. Virtual environments will be considered either virtual reality (VR) or augmented reality (AR) by their virtual nature. Even if AR is usually considered an extension of VR, since most of the augmentations of reality are computer graphics, the nature of AR is also virtual and will be taken as a virtual environment. The applications are divided in two main categories, training and assistance. Each category has subsections for the use of haptic systems in virtual environments in education, medicine, and industry. Finally, an alternative category of entertainment is also discussed. Some repre- sentative research on each area of application isPROOF described to analyze and to discuss which are the trends and challenges related to the applications of haptic systems in virtual environments. Keywords Haptic systems Á Virtual environments Á Augmented reality Virtual reality 13.1 Introduction Humans are in constant interaction with different environments. The interaction is possible through the sense of sight, touch, taste, hearing, and smell. Basic knowledgeREVISED is learned through the five senses. Particularly, sight, hearing, and touch are essential to acquire knowledge based on the fact that people learn best when they see, hear, and do (Volunteer Development 4H-CLUB-100 2016).
    [Show full text]
  • Download File
    SUPPORTING MULTI-USER INTERACTION IN CO-LOCATED AND REMOTE AUGMENTED REALITY BY IMPROVING REFERENCE PERFORMANCE AND DECREASING PHYSICAL INTERFERENCE Ohan Oda Submitted in partial fulfillment of the requirements for the degree of Doctor of Philosophy in the Graduate School of Arts and Sciences COLUMBIA UNIVERSITY 2016 © 2016 Ohan Oda All rights reserved ABSTRACT Supporting Multi-User Interaction in Co-Located and Remote Augmented Reality by Improving Reference Performance and Decreasing Physical Interference Ohan Oda One of the most fundamental components of our daily lives is social interaction, ranging from simple activities, such as purchasing a donut in a bakery on the way to work, to complex ones, such as instructing a remote colleague how to repair a broken automobile. While we inter- act with others, various challenges may arise, such as miscommunication or physical interference. In a bakery, a clerk may misunderstand the donut at which a customer was pointing due to the uncertainty of their finger direction. In a repair task, a technician may remove the wrong bolt and accidentally hit another user while replacing broken parts due to unclear instructions and lack of attention while communicating with a remote advisor. This dissertation explores techniques for supporting multi-user 3D interaction in aug- mented reality in a way that addresses these challenges. Augmented Reality (AR) refers to inter- actively overlaying geometrically registered virtual media on the real world. In particular, we address how an AR system can use overlaid graphics to assist users in referencing local objects accurately and remote objects efficiently, and prevent co-located users from physically interfer- ing with each other.
    [Show full text]
  • Virtual Fixtures in VR – Perceptual Overlays for Assisted Teleoperation, Teleprogramming and Learning
    International Conference on Artificial Reality and Telexistence Eurographics Symposium on Virtual Environments (2018) G. Bruder, S. Cobb, and S. Yoshimoto (Editors) Virtual Fixtures in VR – Perceptual Overlays for Assisted Teleoperation, Teleprogramming and Learning D. Krupke1;2 , J. Zhang2 and F. Steinicke1 1University of Hamburg, HCI, Germany 2University of Hamburg, TAMS, Germany Figure 1: The red line indicates a recorded trajectory of a single trial during the experiment, where the gripper is teleoperated to grasp the blue sphere on the left and to move it to the green goal zone on the right precisely and fast. In the experiment effects of visual and haptic assistance at difficult passages are examined. Abstract Current advances in mixed reality (MR) technology achieves both, making the sensations more immersive and plausible, but also increase the utilization of these technologies in robotics. Their low-cost and the low effort to integrate such a system in complex facilities makes them interesting for industrial application. We present an efficient implementation of “virtual fixtures” [BR92] and the evaluation in a task of three different difficulties. Finally, it is discussed if the method is successfully implemented without real physical barriers and if human performance is effected in teleoperation or teleprogramming of industrial robots. CCS Concepts • Human-centered computing → Virtual reality; • Computer systems organization → External interfaces for robotics; 1. Introduction system as a master-slave control system and presented images of the remote site to the operator meanwhile the movements of the Building and using tools for enhancing and improving our capa- operator were physically constraint in different ways. Current de- bilities is part of the human nature.
    [Show full text]
  • Web-Based Indoor Positioning System Using QR-Codes As Mark- Ers
    Instructor Giullio Jacucci, Petri Savolainen Web-based indoor positioning system using QR-codes as mark- ers Zhen Shi Helsinki November 5, 2020 Master of Computer Science UNIVERSITY OF HELSINKI Department of Computer Science HELSINGIN YLIOPISTO — HELSINGFORS UNIVERSITET — UNIVERSITY OF HELSINKI Tiedekunta — Fakultet — Faculty Laitos — Institution — Department Faculty of Science Department of Computer Science Tekijä — Författare — Author Zhen Shi Työn nimi — Arbetets titel — Title Web-based indoor positioning system using QR-codes as markers Oppiaine — Läroämne — Subject Computer Science Työn laji — Arbetets art — Level Aika — Datum — Month and year Sivumäärä — Sidoantal — Number of pages Master of Computer Science November 5, 2020 40 pages + 0 appendix pages Tiivistelmä — Referat — Abstract Location tracking has been quite an important tool in our daily life. The outdoor location tracking can easily be supported by GPS. However, the technology of tracking smart device users indoor position is not at the same maturity level as outdoor tracking. AR technology could enable the tracking on users indoor location by scanning the AR marker with their smart devices. However, due to several limitations (capacity, error tolerance, etc.) AR markers are not widely adopted. Therefore, not serving as a good candidate to be a tracking marker. This paper carries out a research question whether QR code can replace the AR marker as the tracking marker to detect smart devices’ user indoor position. The paper has discussed the research question by researching the background of the QR code and AR technology. According to the research, QR code should be a suitable choice to implement as a tracking marker. Comparing to the AR marker, QR code has a better capacity, higher error tolerance, and widely adopted.
    [Show full text]