A Multiple Optical Tracking Based Approach for Enhancing Hand- Based Interaction in Virtual Reality Simulations

Total Page:16

File Type:pdf, Size:1020Kb

A Multiple Optical Tracking Based Approach for Enhancing Hand- Based Interaction in Virtual Reality Simulations A Multiple Optical Tracking Based Approach for Enhancing Hand- Based Interaction in Virtual Reality Simulations Adam Grant Worrallo BSc (Hons), MSc A thesis submitted in partial fulfilment of the requirements of the University of Wolverhampton for the degree of Doctor of Philosophy September 2020 This work or any part thereof has not previously been presented in any form to the University or to any other body whether for the purposes of assessment, publication or for any other purpose (unless otherwise indicated). Save for any express acknowledgments, references and/or bibliographies cited in the work, I confirm that the intellectual content of the work is the result of my own efforts and of no other person. The right of Adam Grant Worrallo to be identified as author of this work is asserted in accordance with ss.77 and 78 of the Copyright, Designs and Patents Act 1988. At this date copyright is owned by the author. Signature……A G Worrallo……………………………. Date…………26/09/2020…………………………… II Acknowledgements Throughout my PhD I have received a great deal of support and advice. I would like to first thank my supervisor, Dr Tom Hartley whose expertise, advice and endless support was invaluable throughout my studies. In addition, I would like to thank my PhD administrator, friends and colleagues at the University of Wolverhampton for their support. I would also like to thank my entire family for their endless support and encouragement throughout my studies. You were always there for me with a sympathetic ear and made the hardest parts seem that much easier. I would like to say a special thank you to my uncles; Dave and Tim, you lit and nurtured my passion for computer science in my early years. I would not be where I am today without your input. Finally, I need to say a special thank you to my father, Mark, for always being there for me no matter what. From my first day of school, you have always supported me with my studies, going to the library with me on a Saturday morning and helping with homework in the afternoon. It’s been many years since we did homework together, but your support has never wavered. Your endless support during this endeavour has been invaluable. Whether it was offering advice or just a friendly face at the end of a long day, I could not have done this without you. III Abstract Research exploring natural virtual reality interaction has seen significant success in optical tracker-based approaches, enabling users to freely interact using their hands. Optical based trackers can provide users with real-time, high-fidelity virtual hand representations for natural interaction and an immersive experience. However, work in this area has identified four issues: occlusion, field-of-view, stability and accuracy. To overcome the four key issues, researchers have investigated approaches such as using multiple sensors. Research has shown multi-sensor-based approaches to be effective in improving recognition accuracy. However, such approaches typically use statically positioned sensors, which introduce body occlusion issues that make tracking hands challenging. Machine learning approaches have also been explored to improve gesture recognition. However, such approaches typically require a pre-set gesture vocabulary limiting user actions with larger vocabularies hindering real-time performance. This thesis presents an optical hand-based interaction system that comprises two Leap Motion sensors mounted onto a VR headset at different orientations. Novel approaches to the aggregation and validation of sensor data are presented. A machine learning sub-system is developed to validate hand data received by the sensors. Occlusion detection, stability detection, inferred hands and a hand interpolation sub-system are also developed to ensure that valid hand representations are always shown to the user. In addition, a mesh conformation sub-system ensures 3D objects are appropriately held in a user’s IV virtual hand. The presented system addresses the four key issues of optical sessions to provide a smooth and consistent user experience. The MOT system is evaluated against traditional interaction approaches; gloves, motion controllers and a single front-facing sensor configuration. The comparative sensor evaluation analysed the validity and availability of tracking data, along with each sensors effect on the MOT system. The results show the MOT provides a more stable experience than the front-facing configuration and produces significantly more valid tracking data. The results also demonstrated the effectiveness of a 45-degree sensor configuration in comparison to a front-facing. Furthermore, the results demonstrated the effectiveness of the MOT systems solutions at handling the four key issues with optical trackers. V Contents Acknowledgements ...................................................................................... III Abstract ....................................................................................................... IV List of Figures ............................................................................................. XII List of Equations ........................................................................................ XXI List of Algorithms ...................................................................................... XXII List of Tables ........................................................................................... XXIII List of Abbreviations ............................................................................... XXVI Chapter 1 – Introduction ............................................................................... 1 1.1 – Virtual Reality ................................................................................... 1 1.2 – Serious Games ................................................................................. 5 1.3 – Hand Interaction ............................................................................... 6 1.4 – Optical Sensor Issues ....................................................................... 8 1.5 – Academic Questions, Aims and Objectives of the Research ........... 14 1.6 – Thesis Contributions ....................................................................... 16 1.7 – Research Methodology ................................................................... 17 1.8 – Thesis Organization ........................................................................ 21 Chapter 2 – Literature Review .................................................................... 22 2.1 – Overview of Interaction Techniques/Methods used in Virtual Reality ................................................................................................................ 24 2.2 – Hand-Based Input Mechanisms ...................................................... 30 VI 2.2.1 – Indirect/Contact Based Interaction Approaches in VR .............. 31 2.2.2 – Vision/Optical Based Interaction Approaches in VR ................. 48 2.3 – Machine Learning in Optical Tracking Systems .............................. 62 2.4 – Immersion & Presence in VR .......................................................... 68 2.5 – Interactive Virtual Reality in Education ............................................ 73 2.6 – Summary ........................................................................................ 80 Chapter 3 – Design Phase One .................................................................. 84 3.1 – Custom Bracket & External Sensor Configuration ........................... 87 3.2 – External Data Packet Handler ......................................................... 91 3.3 – Data Processor ............................................................................... 95 3.4 – Data Aggregation System ............................................................... 96 3.5 – Hand Validation System ............................................................... 103 3.6 – Post Aggregation Processing System ........................................... 109 3.7 – Vive Arm Tracking System............................................................ 110 3.8 – Inferred Hand Pose System .......................................................... 112 3.9 – Mesh Conformation System .......................................................... 118 3.10 – Summary .................................................................................... 128 Chapter 4 – Experimentation Phase One .................................................. 129 4.1 – Experimental Setup & Procedure .................................................. 130 4.2 – Semi-structured Interview Analysis ............................................... 138 4.2.1 – Likert Questionnaire Analysis ................................................. 138 VII 4.2.2 - Thematic Analysis ................................................................... 141 4.3 – Tracking Data Validity ................................................................... 146 4.3.1 – Data Validation Process ......................................................... 147 4.3.2 – Sensor Tracking Data Validity ................................................ 149 4.3.3 – Chi-Square Test ..................................................................... 153 4.3.4 – Linear Regression Test .......................................................... 155 4.4 – User Performance ........................................................................ 157 4.4.1 - Total Completion Time Condition Comparison ........................ 161 4.4.2
Recommended publications
  • Evaluating Performance Benefits of Head Tracking in Modern Video
    Evaluating Performance Benefits of Head Tracking in Modern Video Games Arun Kulshreshth Joseph J. LaViola Jr. Department of EECS Department of EECS University of Central Florida University of Central Florida 4000 Central Florida Blvd 4000 Central Florida Blvd Orlando, FL 32816, USA Orlando, FL 32816, USA [email protected] [email protected] ABSTRACT PlayStation Move, TrackIR 5) that support 3D spatial in- teraction have been implemented and made available to con- We present a study that investigates user performance ben- sumers. Head tracking is one example of an interaction tech- efits of using head tracking in modern video games. We nique, commonly used in the virtual and augmented reality explored four di↵erent carefully chosen commercial games communities [2, 7, 9], that has potential to be a useful ap- with tasks which can potentially benefit from head tracking. proach for controlling certain gaming tasks. Recent work on For each game, quantitative and qualitative measures were head tracking and video games has shown some potential taken to determine if users performed better and learned for this type of gaming interface. For example, Sko et al. faster in the experimental group (with head tracking) than [10] proposed a taxonomy of head gestures for first person in the control group (without head tracking). A game ex- shooter (FPS) games and showed that some of their tech- pertise pre-questionnaire was used to classify participants niques (peering, zooming, iron-sighting and spinning) are into casual and expert categories to analyze a possible im- useful in games. In addition, previous studies [13, 14] have pact on performance di↵erences.
    [Show full text]
  • Inverse Kinematic Infrared Optical Finger Tracking
    Inverse Kinematic Infrared Optical Finger Tracking Gerrit Hillebrand1, Martin Bauer2, Kurt Achatz1, Gudrun Klinker2 1Advanced Realtime Tracking GmbH ∗ Am Oferl¨ 3 Weilheim, Germany 2Technische Universit¨atM¨unchen Institut f¨urInformatik Boltzmannstr. 3, 85748 Garching b. M¨unchen, Germany e·mail: [email protected] Abstract the fingertips are tracked. All further parameters, such as the angles between the phalanx bones, are Human hand and finger tracking is an important calculated from that. This approach is inverse input modality for both virtual and augmented compared to the current glove technology and re- reality. We present a novel device that overcomes sults in much better accuracy. Additionally, the the disadvantages of current glove-based or vision- device is lightweight and comfortable to use. based solutions by using inverse-kinematic models of the human hand. The fingertips are tracked by an optical infrared tracking system and the pose of the phalanxes is calculated from the known anatomy of the hand. The new device is lightweight and accurate and allows almost arbitrary hand movement. Exami- nations of the flexibility of the hand have shown that this new approach is practical because ambi- guities in the relationship between finger tip posi- tions and joint positions, which are theoretically possible, occur only scarcely in practical use. Figure 1: The inverse kinematic infrared optical finger tracker Keywords: Optical Tracking, Finger Tracking, Motion Capture 1 Introduction 1.1 Related Work While even a standard keyboard can be seen For many people, the most important and natural as an input device based on human finger mo- way to interact with the environment is by using tion, we consider only input methods that allow their hands.
    [Show full text]
  • Motion Tracking in Field Sports Using GPS And
    Motion tracking in field sports using GPS and IMU M. Roobeek Master of Science Thesis Delft Center for Systems and Control Motion tracking in field sports using GPS and IMU Master of Science Thesis For the degree of Master of Science in Systems & Control at Delft University of Technology M. Roobeek February 16, 2017 Faculty of Mechanical, Maritime and Materials Engineering (3mE) · Delft University of Technology The work in this thesis was conducted on behalf of, and carried out in cooperation with JOHAN Sports. Their cooperation is hereby gratefully acknowledged. Copyright c Delft Center for Systems and Control (DCSC) All rights reserved. Abstract Injuries are a sportsman’s worst nightmare: They withhold players from playing, they weaken teams and they force clubs to buy a bench-full of substitutes. On top of that, injuries can cause an awful lot of pain. Regulating match and training intensity can drastically decrease the risk of getting injured. Furthermore, it can optimally prepare players for matchday. To enable player load regulation, accurate measurements of player positions, velocities and especially accelerations are required. To measure these quantities, JOHAN Sports develops a sports player motion tracking system. The device that is used for motion tracking contains a 9-DoF MEMS Inertial Measurement Unit (IMU) and a GPS receiver. These low-cost MEMS sensors are combined via sensor fusion. The challenge in this filtering problem lies in the limited and low-quality sensor measurements combined with the high-dynamic player motion consisting of rapid orientation changes and sensed impacts in every step. Being able to overcome these challenges paves the way for injury prevention, saving sports clubs, teams and players a lot of misery.
    [Show full text]
  • Augmented and Virtual Reality in Operations
    Augmented and Virtual Reality in Operations A guide for investment Introduction Augmented Reality and Virtual Reality (AR/VR) are not new, but recent advances in computational power, storage, graphics processing, and high-resolution displays have helped overcome some of the constraints that have stood in the way of the widespread use of these immersive technologies. Global spending on AR/VR is now expected to reach $17.8 billion in 2018, an increase of nearly 95% over the $9.1 billion estimate for 2017.1 Given this impetus, we wanted to understand more about the current state of play with AR/VR in the enterprise. We surveyed executives at 700+ companies, including 600 companies who are either experimenting with or implementing AR/VR; spoke with AR/VR-focused leadership at global companies, start-ups, and vendors; and analyzed over 35 use cases, with a particular focus on: • The use of AR/VR for internal industrial company operations, including, but not limited to, design, engineering, and field services • The automotive, utility, and discrete manufacturing sectors. This research focuses on organizations that have initiated their AR/VR journey, whether by experimentation or implementation, and aims to answer a number of questions: • How is AR/VR being used today? • Where should organizations invest – what are the use cases that generate the most value? • How can companies either begin or evolve their initiatives? Outside of this report, terms you will commonly find when reading about AR/VR are mixed reality, hybrid reality, or merged reality in which both the physical and the virtual or digital realities coexist or even merge.
    [Show full text]
  • Airpen: a Touchless Fingertip Based Gestural Interface for Smartphones and Head-Mounted Devices
    AirPen: A Touchless Fingertip Based Gestural Interface for Smartphones and Head-Mounted Devices Varun Jain Ramya Hebbalaguppe [email protected] [email protected] TCS Research TCS Research New Delhi, India New Delhi, India ABSTRACT Hand gestures are an intuitive, socially acceptable, and a non-intrusive interaction modality in Mixed Reality (MR) and smartphone based applications. Unlike speech interfaces, they tend to perform well even in shared and public spaces. Hand gestures can also be used to interact with smartphones in situations where the user’s ability to physically touch the device is impaired. However, accurate gesture recognition can be achieved through state-of-the-art deep learning models or with the use of expensive sensors. Despite the robustness of these deep learning models, they are computationally heavy and memory hungry, and obtaining real-time performance on-device without additional hardware is still a challenge. To address this, we propose AirPen: an analogue to pen on paper, but in air, for in-air writing and gestural commands that works seamlessly in First and Second Person View. The models are trained on a GPU machine and ported on an Android smartphone. AirPen comprises of three deep learning models that work in tandem: MobileNetV2 for hand localisation, our custom fingertip regression architecture followed by a Bi-LSTM model for gesture classification. The overall framework works in real-time on mobile devices and achieves a classification accuracy of 80% with an average latency of only 0:12 s. Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page.
    [Show full text]
  • Virtual Reality 3D Space & Tracking
    Virtual Reality 3D Space & Tracking Ján Cíger Reviatech SAS [email protected] http://janoc.rd-h.com/ 16.10.2014 Ján Cíger (Reviatech SAS) RV01 16.10.2014 1 / 57 Outline Who am I? Tracking Coordinates in 3D Space Introduction Position in 3D Technologies Orientation in 3D VRPN Euler angles What is VRPN Axis/angle representation Simple Example Rotation matrices (3 × 3) Quaternions Resources Ján Cíger (Reviatech SAS) RV01 16.10.2014 2 / 57 Who am I? Who am I? PhD from VRlab, EPFL, Switzerland I AI for virtual humans, how to make autonomous characters ("NPCs") smarter I Lab focused on human character animation, motion capture, interaction in VR I ≈ 20 PhD students, several post-docs, many master students. VRlab: http://vrlab.epfl.ch IIG: http://iig.epfl.ch Ján Cíger (Reviatech SAS) RV01 16.10.2014 3 / 57 Who am I? Who am I? SensoramaLab, Aalborg University, Denmark I Building a new lab focusing on new media, interaction, use in rehabilitation I Large projection system, tracking, many interactive demos I Teaching bachelor & master students, project supervision. Ján Cíger (Reviatech SAS) RV01 16.10.2014 4 / 57 Who am I? Who am I? Reviatech SAS I Applied research, looking for new technologies useful for our products/customers I Tracking, augmented reality, AI, building software & hardware prototypes. http://www.reviatech.com/ Ján Cíger (Reviatech SAS) RV01 16.10.2014 5 / 57 Coordinates in 3D Space Position in 3D Outline Who am I? Tracking Coordinates in 3D Space Introduction Position in 3D Technologies Orientation in 3D VRPN Euler angles What is VRPN Axis/angle representation Simple Example Rotation matrices (3 × 3) Quaternions Resources Ján Cíger (Reviatech SAS) RV01 16.10.2014 6 / 57 Coordinates in 3D Space Position in 3D 3D Position Carthesian I 3 perpendicular axes I Left or right handed! I Z-up vs Y-up! I Usually same scale on all axes http://viz.aset.psu.edu/gho/sem_notes/3d_ I Common units – meters, fundamentals/html/3d_coordinates.html millimetres, inches, feet.
    [Show full text]
  • Bakalářská Práce Software Pro Rehabilitaci Paže Ve Virtuální Realitě
    Západočeská univerzita v Plzni Fakulta aplikovaných věd Katedra informatiky a výpočetní techniky Bakalářská práce Software pro rehabilitaci paže ve virtuální realitě Plzeň 2020 Jakub Kodera Místo této strany bude zadání práce. Prohlášení Prohlašuji, že jsem bakalářskou práci vypracoval samostatně a výhradně s použitím citovaných pramenů. V Plzni dne 7. května 2020 Jakub Kodera Abstract Arm rehabilitation software in virtual reality. The work is a part of a larger project, that is trying to achieve rehabilitation of the upper limb, in pa- tients suffering from a neurological disorder, in virtual reality. The result of this work is an application, that allows patients to perform rehabilitation exercises. The technologies used to develop the software are discussed here, mainly the HTC Vive system and the Unity game engine. The application is designed so that it can be further expanded. The application provides patients with a vizualization of the movement, that they are to perform. The therapeut obtains data and statistics of the exercises performed by the patient. Abstrakt Práce je součástí projektu, který se zabývá rehabilitací horní končetiny, u pacientů trpících neurologickou poruchou, ve virtuální realitě. Výsledkem této práce je aplikace umožňující provádět rehabilitačních cvičení. Jsou zde probírány technologie použité při vývoji softwaru. Zejména systém HTC Vive a herní engine Unity. Aplikace je navržena tak, aby ji bylo možné dále rozšiřovat. Aplikace poskytuje pacientům vizualizaci prováděného pohybu. Terapeut aplikací získá data a statistiky o cvičeních vykonaných pacientem. Obsah 1 Úvod7 2 Roztroušená skleróza mozkomíšní8 2.1 Průběh nemoci.........................8 2.1.1 Typy roztroušené sklerózy............... 10 2.2 Symptomy............................ 11 2.3 Léčba.............................
    [Show full text]
  • Auraring: Precise Electromagnetic Finger Tracking
    AuraRing: Precise Electromagnetic Finger Tracking FARSHID SALEMI PARIZI∗, ERIC WHITMIRE∗, and SHWETAK PATEL, University of Washington Fig. 1. AuraRing is 5-DoF electromagnetic tracker that enables precise, accurate, and fine-grained finger tracking for AR, VR and wearable applications. Left: A user writes the word "hello" in the air. Right: Using AuraRing to play a songinamusic application on a smart glass. Wearable computing platforms, such as smartwatches and head-mounted mixed reality displays, demand new input devices for high-fidelity interaction. We present AuraRing, a wearable magnetic tracking system designed for tracking fine-grained finger movement. The hardware consists of a ring with an embedded electromagnetic transmitter coil and a wristband with multiple sensor coils. By measuring the magnetic fields at different points around the wrist, AuraRing estimates thefive degree-of-freedom pose of the ring. We develop two different approaches to pose reconstruction—a first-principles iterative approach and a closed-form neural network approach. Notably, AuraRing requires no runtime supervised training, ensuring user and session independence. AuraRing has a resolution of 0:1 mm and a dynamic accuracy of 4:4 mm, as measured through a user evaluation with optical ground truth. The ring is completely self-contained and consumes just 2:3 mW of power. CCS Concepts: • Human-centered computing → Interaction devices; Ubiquitous and mobile computing. Additional Key Words and Phrases: Electromagnetic tracking, mixed reality, wearable, input, finger tracking ACM Reference Format: Farshid Salemi Parizi, Eric Whitmire, and Shwetak Patel. 2019. AuraRing: Precise Electromagnetic Finger Tracking. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 3, 4, Article 150 (December 2019), 28 pages.
    [Show full text]
  • Virtual & Augmented Reality
    EXCERPTED FROM THE ORIGINAL: See inside cover for details. FINANCIAL BACKING GLOBAL INTEREST SHIPPING OUT SELLING OUT INTEREST IN THE PAST READY TO BUILD A WAVE OF CONTENT ON THE WAY RETAIL VALUE HOME REDESIGN, REIMAGINED EASIER TO IMAGINE YOURSELF AT HOME Video • Jaunt The Ecosystem • NextVR Virtual Reality / Augmented Reality • VRSE • Oculus Story Studio • GoPro • IG Port Processors Games • TI Applications • Sony • Qualcomm • Ubisoft 3D Audio • STMicro Graphics • CCP Games • Nvidia • TI • Realtek • Oculus Story Studio • AMD • Wolfson • Himax • Tammeka Games • Qualcomm • Realtek • MediaTek • Pixel Titans • Intel • Capcom Augmented Reality Virtual Reality Engineering • Microsoft HoloLens • Facebook Oculus Head–mounted devices • Autodesk • Google Glass • Samsung Gear VR • Dassault Systèmes • Magic Leap • Google Cardboard • IrisVR • Atheer • HTC Vive • Visidraft • Osterhout Design • Sony PSVR • MakeVR Memory Group • Vuzix iWear (DRAM/SSD) • VR Union Claire • Micron Healthcare • Samsung • Psious • SK Hynix • zSpace • Toshiba • Conquer Mobile • 3D Systems Social • Altspace VR • High Fidelity • Podrift Commerce • Sixense {shopping} • Matterport {real estate} Display • Samsung • JDI Cameras • Himax • 360Heros • Crystal • GoPro Odyssey 3D Lenses • Nokia OZO • Wearality • Jaunt NEO • Zeiss • Matterport Pro 3D Components • Canon • Nikon Haptics • Largan • Alps Position/ Room Tracker • AAC • Hon Hai • Nidec • Pegatron • Flex • Jabil • HTC Motion Sensors • Leap Motion • InvenSense • TI • STMicro • Honeywell The Ecosystem Virtual Reality / Augmented
    [Show full text]
  • Lightweight Palm and Finger Tracking for Real-Time 3D Gesture Control
    Lightweight Palm and Finger Tracking for Real-Time 3D Gesture Control Georg Hackenberg* Rod McCall‡ Wolfgang BrollD Fraunhofer FIT Fraunhofer FIT Fraunhofer FIT Freehand multi-touch has been explored within various ABSTRACT approaches, e.g. Oblong’s g-speak1, after initially having been We present a novel technique implementing barehanded introduced in the popular Hollywood movie “Minority Report”. interaction with virtual 3D content by employing a time-of-flight They usually depended heavily on hand-worn gloves, markers, camera. The system improves on existing 3D multi-touch systems wrists, or other input devices, and typically did not achieve the by working regardless of lighting conditions and supplying a intuitiveness, simplicity and efficiency of surface (2D) based working volume large enough for multiple users. Previous multi-touch techniques2. In contrast to those approaches, the goal systems were limited either by environmental requirements, of our approach was to use barehanded interaction as a working volume, or computational resources necessary for real- replacement for surface based interaction. time operation. By employing a time-of-flight camera, the system Only a vision-based approach will allow for freehand and is capable of reliably recognizing gestures at the finger level in barehanded 3D multi-touch interaction. The system must also real-time at more than 50 fps with commodity computer hardware provide sufficient solutions for the following four steps: detection using our newly developed precision hand and finger-tracking of hand position without prior knowledge of existence; for each algorithm. Building on this algorithm, the system performs appearance determine pose from image cues; track appearances gesture recognition with simple constraint modeling over over time; recognize gestures based on their trajectory and pose statistical aggregations of the hand appearances in a working information.
    [Show full text]
  • Virtual Reality and Augmented Reality a Survey from Scania’S Perspective
    EXAMENSARBETE INOM MASKINTEKNIK, AVANCERAD NIVÅ, 30 HP STOCKHOLM, SVERIGE 2019 Virtual Reality and Augmented Reality A Survey from Scania’s Perspective KARL INGERSTAM KTH SKOLAN FÖR INDUSTRIELL TEKNIK OCH MANAGEMENT Virtual Reality and Augmented Reality A Survey from Scania’s Perspective Karl Ingerstam 2019 Master of Science Thesis TPRMM 2019 KTH – Industrial Engineering and Management Production Engineering SE-100 44 Stockholm Abstract Virtual reality and augmented reality are technological fields that have developed and expanded at a great pace the last few years. A virtual reality is a digitally created environment where computer- generated elements are displayed to a user via different interfaces for the respective senses. Video is used for displaying images, creating a realistic environment, while audio is played to stimulate hearing and other sorts of feedback is used to stimulate the sense of touch in particular. Augmented reality is a sub-category of virtual reality where the user sees the real surroundings, but computer-generated imagery is displayed on top of objects in the environment. This type of technology brings a lot of new possibilities and potential use cases in all sorts of areas, ranging from personal entertainment, communication and education to medicine and heavy industry. Scania is a global manufacturer of heavy trucks and buses, and provider of related services, based in Sweden. By studying Scania’s different departments and surveying the fields of virtual reality and augmented reality, the aim of this thesis is to identify situations and use cases where there is potential for Scania to implement virtual reality and augmented reality technology. This thesis also studies what obstacles implementation of these technologies bring.
    [Show full text]
  • Perception-Action Loop in the Experience of Virtual Environments
    Perception-Action Loop in the Experience of Virtual Environments by Seung Wook Kim A dissertation submitted in partial satisfaction of the requirements for the degree of Doctor of Philosophy in Architecture and the Designated Emphasis in New Media in the Graduate Division of the University of California, Berkeley Committee in charge: Professor Yehuda E. Kalay, Chair Professor Galen Cranz Professor John F. Canny Fall 2009 Abstract Perception-Action Loop in the Experience of Virtual Environments by Seung Wook Kim Doctor of Philosophy in Architecture and the Designated Emphasis in New Media University of California, Berkeley Professor Yehuda Kalay, Chair The goal of this work is to develop an approach to the design of natural and immersive interaction methods for three-dimensional virtual environments. My thesis is that habitation and presence in any environment are based on a continuous process of perception and action between a person and his/her surroundings. The current practice of virtual environments, however, disconnects this intrinsic loop, separating perception and action into two different ‘worlds’—a physical one (for perception) and a virtual one (for action). This research is aimed at bridging the gap between those two worlds. Being drawn from perceptual philosophy and psychology, the theoretical study in this dissertation identifies three embodiments of natural perception-action loop: direct perceptual acts, proprioceptive locomotion, and motor intentionality. These concepts form the basis for the interaction methods proposed in this work, and I demonstrate these methods by implementing pertinent prototype systems. First, I suggest a view-dependent, non-planar display space that supports natural perceptual actions, thereby enhancing our field of view as well as depth perception.
    [Show full text]