ECEN 5623 RT Embedded Systems

Total Page:16

File Type:pdf, Size:1020Kb

ECEN 5623 RT Embedded Systems CS415 Human Computer Interaction Lecture 12 – Ubiquitous Computing VR, AR, IoT, Pervasive (Ch. 20) November 11, 2019 Sam Siewert Remaining Assignments Assignment #5 – Explore HCI’s, Propose Group Project (Groups of 2, 3 or 4) Assignment #6 – Proof-of-Concept or Evaluation with Final Report Final Oral Exam – Presentation of Project FINAL: 10:15-12:15pm, Sat, Dec 7, Sims&Games, Bldg 61 – Oral Presentation of Assignment #6 Report – With Updates for any Progress since Sam Siewert 2 Ubiquitous Computing Term Credited to Mark Weiser – “The Computer of the 21st Century”, Scientific American, 265(3):66-75, September 1991 – Vision Described Previously at Many HCI Research Organizations [e.g. Xerox PARC, MIT Media Lab, GIT, …] Some Theorized Enabling Technologies (Book) – High speed wireless networks – Computer Vision – Perceptual Computing to Emulate Human Perception and Extend it – Projection, 3D, High Resolution Large Displays and Mobile Computing – Head tracking, Independent Display Eyeware, 3D Glasses – Large Scale Interactive Surfaces (Large Touch Screens) – Data Gloves and Body Suits Sam Siewert 3 Advanced HCI … Future AR/VR and IoT - Daily life integrated computing Microsoft Research HCI Vision Interesting HCI Innovations and Experiments (Examples) – MS Hololens, – Magic Leap, – jibo, – sixthsense, – Oculus VR, – NVIDIA AR/VR, PX, – Intel Real Sense, – Google Project Tango, – Interactive Film, – Gear VR Sam Siewert 4 Ubiquitous Computing [“computing is made to appear anytime and everywhere”] Pranav Mistry, Ted Talk on Interaction (We watched) John Barrett, IoT Ted Talk – Cork Inst. Of Technology – Embedded Systems Research Marco Annunziata, GE, IIoT Ted Talk (Watch …) Privacy in IoT World - Daniel Price TedX (Watch …) Week 12 … IoT and re-visit to AR, VR, Quiz-3 Week 13 … Advanced Avionics and UAV/UAS Week 14 – Quiz-4, Review for Exam #2, A#6 Help Week 15 – Exam #2 Final Oral Exam - 12/14 Sam Siewert 5 IoT and IIoT Reality and Vision Complex machines and systems self- report (digital twin) Built-into devices and systems Secure wireless connectivity Quote from Aviation Week - Jan 2016: Internet of Aircraft Things Last year we saw the next generation of aircraft powered by the capabilities of Internet of Things (IoT) and Big Data hit scales that How much can a have never seen before. What does the future hold? pilot assess H&S At last year’s Paris Air Show for example, Bombardier showcased its C Series jetliner that carries Pratt & Whitney’s Geared Turbo Fan (GTF) from visual engine, which is fitted with 5,000 sensors that generate up to 10 GB of data per second. A single twin-engine aircraft with an average 12-hr. inspection? flight-time can produce up to 844 TB of data. Sam Siewert 6 Markets for Ubiquitous Computing Virtuous link - Games & Animation, HPC – Clusters (ILM, Pixar, Dreamworks, …) – GPU and GP-GPU (NVIDIA, ATI, Intel) – Serious Gaming (Education, Therapy, …) Transportation (Industrials) Nov, 2018 Nov, 2019 – Move people, animals, goods and commodities – Modes - air, rail, road, water, cable, pipeline, space – UAS – aerial transportation at lower cost? Mobile, Internet – Self-Driving Cars Games Energy & Utilities (IIoT) – Smart GRID – generation and distribution – Alternative Energy Sources and storage Cyberphysical – Clean Nuclear and Coal (e.g. TeraPower) – Carbon sequestration Enterprise Services (IT and Telecomm) Cyberphysical – Mobile and Wireless for IoT, IIoT (5th Generation) Cyberphysical Enterprise Health Care (IoT) – Elder care (at home) – Therapeutics at home – Patient monitoring, Health and Fitness monitoring – Big data analytics – pathology, disease Cyberphysical cures, control and prevention Sam Siewert 7 Definitions Ubiquitous Computing – Merge computational artifacts smoothly with world of physical artifacts (applications integrated with things) – Superimpose relevant information over top of the object – Objects in the physical world to manipulate virtual (graspable) – Context aware applications (Who, What, Where, When, Why) – Recommendation Engines, Adware, Personal Assistants – Vannevar Bush (Memex) – Often compared to Hypertext Today VR (Virtual Reality) – Immersive – Desktop – Simulated World AR (Augmented Reality) – Project Onto World – See World Through Mobile Screen / Glasses With Annotation – Requires Computer Vision (Object Recognition) for True AR Sam Siewert 8 Internet of Things A new form of Ubiquitous Computing? – IPv6 - 128 bit address, 2128 , or about 3.4 x 1038 – Eddington_number - Theorized bound on Elements in Universe 256 79 – NEdd = 136×2 , or about 1.57×10 Build processing into industry and everyday life – IIoT – Industrial Internet of Things E.g. Honeywell – IIoT, Intel IIoT, GE IoT – IoT – Everyday and Workday Internet of Things E.g. Intel IoT, IBM IoT – Network infrastructure for IoT (Wireless) E.g. ATT IoT, Verizon IoT, Cisco IoT Security Challenges (Cyberphysical) – Classic Risk, Benefit, Impact, Cost to Mitigate Design Trade-offs – Remote Monitoring and Proactive Maintenance Sam Siewert 9 IoT Processors - Sensors + Wireless Very Small, Wireless (BLE), Wired Ethernet, Built-in Sensors, Very Low-Power E.g. Intel’s Quark, NXP, Texas Instruments TM4C, CC32XX Arduino – Education, Makers and Quick Prototypes Zephyr (Linux Foundation) – Micro / Nano kernel – Boards supported – Paper on Curie/Quark – Example Zephyr IoT Code Simple Executives – E.g. Tiva with PDL Sam Siewert 10 NSF View of IoT Benefits and Risks of the “Internet of Things” Embedded Computing – Cyberphysical Systems, Extension of Embedded Systems (Networked) https://ptolemy.berkeley.edu/projects/cps/ Sam Siewert 11 Embedded Software and IoT Simple Executive on bare metal - TI PDL with TIVA (Microprocessor class/lab - CEC 320/322) Micro or Nano kernel purpose-built for IoT – Linux Foundation Zephyr – FreeRTOS – 8 More - Riot OS, Win10-IoT, VxWorks IIoT, Google Android Things, ARM Mbed, Apple Homekit, Nucleus, Green Hills RTOS or OS on IoT device (CEC 450) – E.g. Raspberry Pi 3b+ (Drone Net Acoustic Tripwire Demonstration) Battery power (power bank) WiFi (802.11) Sensor interfaces (GPS, microphone, IMU) – General info on R-Pi 3b+ for IoT InformationWeek - 10 Raspberry Pi Projects For Learning IoT https://www.raspberrypi.org/blog/getting-started-with-iot/ Sam Siewert 12 Zephry and Linux for IoT Linux Foundation Real-Time IoT Nano / Micro-kernel Sam Siewert 13 What is IoT? Internet of Things (paper on Zephyr / Curie-Quark) – Similar to older idea of “Sensor Networks”, but high bandwidth Internet sensors and actuators – https://elinux.org/RPi_Hub (R Pi is large IoT Device) – Nodes MUST have at least local/wide area wireless interface BLE - Bluetooth Low Energy - mobile phone or tablet 802.11 - Wireless Ethernet 4G LTE or 5G Cellular Security is a major challenge Opportunity is smart devices and “digital twins” 4G LTE, 5G – E.g. Pilot walk-around with tablet for aircraft health and status – Telemetry to manufacturer for predictive maintenance – Simple configuration, provisioning, upgrades Zephyr board support (ARM NXP, Intel, …, TI TIVA?) Xbee, Zigbee, etc. BLE Sam Siewert 14 More IoT Study Digital Circuits - CEC220, CEC222 Embedded Systems - Micro class (CEC320) and lab (CEC322) Undergraduate research with ICARUS (Siewert, Bruder, Andalibi) or Capstone IoT, or Dr. “A”, Dr. Post, etc. Graduate programs – CU ESE program – MIT Sloan IoT business program – Stanford IoT Certificate program Sam Siewert 15 Required Software Basic Nano / Micro Kernel Network stack software – FreeRTOS – BLE network stack – Zephyr – 802.11 network stack – RTEMS – Cellular modem App interfaces – AOS, iOS L7 SCP Service SCP Client L6 Semi-structured file Semi-structured file L5 Sockets API Sockets API L4 TCP TCP L3 IP IP Router/Switch IP L2 IEEE 802.3z - GigE 802.11abgn - Wireless L1 Cat-6 UTP Cat-6 UTP Internet Sam Siewert 16 Ubiquitous Computing Attribute Model IoT systems have more applications than CPS or ES? Relation to Cyberphysical Systems (CPS larger scope than ES?) Relation to Embedded Systems (IoT is a class of ES?) Ubiquitous Computing IoT (context aware, merged physical and virtual) Architecture Class model or IoT (sensors, wireless, processing, digital twin) IoT a superset of ES+CPS? Might be too or Cyberphysical Systems (networked Constraining or Killer app M2M, HCI, secure) Oversimplified (making ES more E.g. ES with attributes prevalent) not found in a CPS or Embedded Systems IoT (i.e. not a proper subset) Sam Siewert 17 IoT Architecture - Class Model ES Less restrictive than attributes = {low power, superset/subset purpose built, specific model services, custom I/O} ES as a base class CPS Transportation ES attributes = attributes CPS inherits from ES {ES’ + networked, {ES’ + durability, security, M2M, HCI} EMI/EMC} IoT inherits form CPS Healthcare CPS IoT attributes = Goal is ubiquitous attributes = {CPS’ + therapeutics, computing {CPS’ + wireless, diagnostics, monitoring} digital twin} Sam Siewert 18 Smart Sensors - Machine Learning with IoT - Neuromorphic Architectures NVIDIA VOLTA GP-GPU Co-Proc NVIDIA Volta - Deep Neural Nets, GP-GPU Google TPU - Tensor processing Unit Google TPU Co-Proc (Cloud) Intel Nervana - NNP (Neural Net Processor) Historical Intel Nervana Chip Neuromorphic - Intel ETANN and Ni1000 Sam Siewert 19 Playing with Concepts Cost Reduction is Key to Become Ubiquitous (Killer App) MERL DiamondTouch (2003) - Book Example Computer Vision Chess, Automatic Duck Hunter Industrial HCI Research MERL Xerox PARC Intel Perceptual Computing MS Research IBM HCI Research Sam Siewert 20 VR, AR Killer
Recommended publications
  • Työohjeistuksen Kehittäminen Lisätyn Todellisuuden Avulla
    Minna Vanhatapio TYÖOHJEISTUKSEN KEHITTÄMINEN LISÄTYN TODELLISUUDEN AVULLA Tukiasematehtaan tuotanto TYÖOHJEISTUKSEN KEHITTÄMINEN LISÄTYN TODELLISUUDEN AVULLA Tukiasematehtaan tuotanto Minna Vanhatapio Opinnäytetyö Syksy 2016 YAMK, Teknologialiiketoiminta Oulun ammattikorkeakoulu TIIVISTELMÄ Oulun ammattikorkeakoulu Ylempi ammattikorkeakoulututkinto, Teknologialiiketoiminta Tekijä: Minna Vanhatapio Opinnäytetyön nimi: Työohjeistuksen kehittäminen lisätyn todellisuuden avulla, Tukiasematehtaan tuotanto Työn ohjaaja: Hannu Päätalo Työnvalmistumislukukausi ja -vuosi: Syksy 2016 Sivumäärä: 107 Opinnäytetyössä tutustuttiin Augmented Reality -tekniikkaan eli lisättyyn todellisuuteen. Lisätyllä todellisuudella tarkoitetaan reaaliaikaista näkymää, jonka päälle lisätään tietokonegrafiikalla tuo- tettua informaatiota kuten 3D-kuvia, ääntä ja videoita. Informaatio voidaan näyttää esimerkiksi äly- puhelimessa, tabletissa, tietokoneen näytöllä tai älylaseilla. Tavoitteena oli tarjota toimeksiantajalle kattava kuva lisätystä todellisuudesta ja sen tämän hetkisistä mahdollisuuksista työohjeistuksessa sekä selvittää mitä hyötyjä sillä voitaisiin saavuttaa Nokia Networksin tukiasematuotannossa. Työssä tutkittiin voitaisiinko lisätyn todellisuuden avulla tuotannon työohjeistusta parantaa, sekä pohdittiin laajemmin mitä tekniikan käyttöönotto vaatii ja mitä kaikkea on otettava huomioon. Tutki- mus suoritettiin tutustumalla tekniikkaa kehittäneiden tutkijoiden tutkimuksiin, käyttäjien ja eri käyt- töalojen ammattilaisten arviointeihin sekä haastateltiin
    [Show full text]
  • Exploratory Research Into Potential Practical Uses Of
    50th ASC Annual International Conference Proceedings Copyright 2014 by the Associated Schools of Construction Exploratory Research into Potential Practical uses of Next Generation Wearable Wireless Voice-Activated Augmented Reality (VAAR) Devices by Building Construction Site Personnel Christopher J. Willis PhD, CAPM, LEED Green Assoc., P.Eng Concordia University Montreal Quebec The miniaturization and increased functionalities of next generation augmented reality (AR) devices, as well as advances in computing technology in the form of cloud computing, is moving the building construction industry closer to adoption of AR devices for use by building construction site personnel. There is therefore a need to understand the potential practical uses of next generation AR devices in building construction site work. A conceptualization of a next generation AR device suitable for use by site personnel is provided. Based on this conceptualization, a focus group of industry professionals and postgraduate researchers have determined that potential practical uses of such a device include: easy access to digital information to support work tasks, live streaming of videos of tasks being worked on, and easy creation of a repository of as-built photographs and videos. Potential applied research studies that will aid in the adoption of next generation AR devices by site personnel include those associated with usability testing, labor productivity measurement and improvement, and suitability testing based on nature of work tasks. The major implication of this exploratory study is that its findings will help to bridge the gap between next generation AR devices and practical use in building construction. Keywords: Augmented Reality, Building Construction, Cloud Computing, Next Generation, Practical Use Introduction Augmented reality (AR) is an emerging technology that is increasingly acquiring greater relevance and usage.
    [Show full text]
  • Augmented Reality Environments for Immersive and Non Linear Virtual Storytelling
    Augmented reality environments for immersive and non linear virtual storytelling Stefanos Kougioumtzis University of Piraeus, Department of Informatics 80, Karaoli & Dimitriou St., 185 34 Piraeus, Greece [email protected] Nikitas N. Karanikolas Technological Educational Institution (TEI) of Athens, Department of Informatics Ag. Spyridonos St., 12210 Aigaleo, Greece [email protected] Themis Panayiotopoulos University of Piraeus, Department of Informatics 80, Karaoli & Dimitriou St., 185 34 Piraeus, Greece [email protected] Abstract The work reported in this paper has the purpose of creating a virtual environment where the recipients (the user) will experience a virtual story telling. In order to introduce the user to a plausible, near to reality experience, we use augmentation methods to create an environment maintaining its physical properties, while three dimensional virtual objects and textures change it for the needs of a real time and controlled scenario. The augmentation is implemented using marker based techniques while the superimposed objects are rendered through the wearable equipment that the user is carrying. In the current implementation we try to blend intelligent non controllable characters that interpret the users’ choices and actions in order to intensify his or her interest for the story, while the real time events are being acted. The user is free to explore, interact and watch the desired events, while others are being carried out in parallel. Keywords: virtual reality, mixed reality, virtual story-telling, wearable systems, computer vision 1. Introduction and problems outline telling, tangible devices, that have already achieved this goal in some extent. As technology moves forward, the needs and purposes of virtuality are getting more and The techniques of Augmented Reality have more clear.
    [Show full text]
  • Field Service Support with Google Glass and Webrtc
    Field Service Support with Google Glass and WebRTC Support av Fälttekniker med Google Glass och WebRTC PATRIK OLDSBERG Examensarbete inom Datorteknik, Grundnivå, 15 hp Handledare på KTH: Reine Bergström Examinator: Ibrahim Orhan TRITA-STH 2014:68 KTH Skolan för Teknik och Hälsa 136 40 Handen, Sverige Abstract The Internet is dramatically changing the way we com- municate, and it is becoming increasingly important for communication services to adapt to context in which they are used. The goal of this thesis was to research how Google Glass and WebRTC can be used to create a communi- cation system tailored for field service support. A prototype was created where an expert is able to provide guidance for a field technician who is wearing Google Glass. A live video feed is sent from Glass to the expert from which the expert can select individual images. When a still image is selected it is displayed to the technician through Glass, and the expert is able to provide instructions using real time annotations. An algorithm that divides the selected image into segments was implemented using WebGL. This made it possible for the expert to highlight objects in the image by clicking on them. The thesis also investigates different options for ac- cessing the hardware video encoder on Google Glass. Sammanfattning Internet har dramatiskt ändrat hur vi kommunicerar, och det blir allt viktigare för kommunikationssystem att kunna anpassa sig till kontexten som de används i. Målet med det här examensarbetet var att undersöka hur Google Glass och WebRTC kan användas för att skapa ett kommunikationssystem som är skräddarsytt för support av fälttekniker.
    [Show full text]
  • Head-Mounted Mixed Reality Projection Display for Games Production and Entertainment
    Pers Ubiquit Comput DOI 10.1007/s00779-015-0847-y ORIGINAL ARTICLE Head-mounted mixed reality projection display for games production and entertainment 1 2 2 3 Daniel Kade • Kaan Aks¸it • Hakan U¨ rey • Og˘uzhan O¨ zcan Received: 20 November 2014 / Accepted: 3 April 2015 Ó Springer-Verlag London 2015 Abstract This research presents a mixed reality (MR) gaming applications as well. Therefore, our MR prototype application that is designed to be usable during a motion could become of special interest because the prototype is capture shoot and supports actors with their task to per- lightweight, allows for freedom of movement and is a low- form. Through our application, we allow seeing and ex- cost, stand-alone mobile system. Moreover, the prototype ploring a digital environment without occluding an actor’s also allows for 3D vision by mounting additional hardware. field of vision. A prototype was built by combining a retroreflective screen covering surrounding walls and a Keywords Head-mounted projection display Á Mixed headband consisting of a laser scanning projector with a reality Á Motion capture Á Laser projector Á Immersive smartphone. Built-in sensors of a smartphone provide environments Á Games production navigation capabilities in the digital world. The presented system was demonstrated in an initially published paper. Here, we extend these research results with our advances 1 Introduction and discuss the potential use of our prototype in gaming and entertainment applications. To explore this potential Entertainment industry products such as video games and use case, we built a gaming application using our MR films are deeply depending on computer-generated imagery prototype and tested it with 45 participants.
    [Show full text]
  • Towards Disappearing User Interfaces for Ubiquitous Computing: Human Enhancement from Sixth Sense to Super Senses
    J Ambient Intell Human Comput DOI 10.1007/s12652-016-0409-9 ORIGINAL RESEARCH Towards disappearing user interfaces for ubiquitous computing: human enhancement from sixth sense to super senses 1 1 Terence K. L. Hui • R. Simon Sherratt Received: 25 March 2016 / Accepted: 5 September 2016 Ó The Author(s) 2016. This article is published with open access at Springerlink.com Abstract The enhancement of human senses electronically 1 Introduction is possible when pervasive computers interact unnoticeably with humans in ubiquitous computing. The design of The growing importance of human computer interaction computer user interfaces towards ‘‘disappearing’’ forces the (HCI) is a result of the widespread deployment of com- interaction with humans using a content rather than a menu puters connecting with humans. Early computer users were driven approach, thus the emerging requirement for huge computer specialists and programmers when the major number of non-technical users interfacing intuitively with design goals for computers were processing-power and billions of computers in the Internet of Things is met. processing-speed during the time a single mainframe Learning to use particular applications in ubiquitous com- computer occupied a whole room. Usability became a puting is either too slow or sometimes impossible so the profound issue when personal computers prevailed in the design of user interfaces must be naturally enough to late 70s since personnel and non-professionals had joined facilitate intuitive human behaviours. Although humans the user group with no prior computer science knowledge from different racial, cultural and ethnic backgrounds own (Shackel 1997). Ubiquitous Computing proposed by Wei- the same physiological sensory system, the perception to ser (1991)as‘‘The Computer for The 21st Century’’ has the same stimuli outside the human bodies can be different.
    [Show full text]
  • Touch-Based Interaction for Mobile Virtual Reality with Clip-On Lenses
    applied sciences Article “Blurry Touch Finger”: Touch-Based Interaction for Mobile Virtual Reality with Clip-on Lenses Youngwon Ryan Kim 1, Suhan Park 2 and Gerard J. Kim 2,* 1 Korea Electronics Technology Institute, Gwangju 61011, Korea; [email protected] 2 Digital Experience Laboratory, Korea University, Seoul 02841, Korea; [email protected] * Correspondence: [email protected] Received: 15 October 2020; Accepted: 5 November 2020; Published: 8 November 2020 Abstract: In this paper, we propose and explore a touch screen based interaction technique, called the “Blurry Touch Finger” for EasyVR, a mobile VR platform with non-isolating flip-on glasses that allows the fingers accessible to the screen. We demonstrate that, with the proposed technique, the user is able to accurately select virtual objects, seen under the lenses, directly with the fingers even though they are blurred and physically block the target object. This is possible owing to the binocular rivalry that renders the fingertips semi-transparent. We carried out a first stage basic evaluation assessing the object selection performance and general usability of Blurry Touch Finger. The study has revealed that, for objects with the screen space sizes greater than about 0.5 cm, the selection performance and usability of the Blurry Touch Finger, as applied in the EasyVR configuration, was comparable to or higher than those with both the conventional head-directed and hand/controller based ray-casting selection methods. However, for smaller sized objects, much below the size of the fingertip, the touch based selection was both less performing and usable due to the usual fat finger problem and difficulty in stereoscopic focus.
    [Show full text]
  • Motion-Based Interaction for Head-Mounted Displays
    Motion-based Interaction for Head- Mounted Displays Thesis submitted in accordance with the requirements of the University of Liverpool for the degree of Doctor in Philosophy By Wenge Xu May 2021 PGR Declaration of Academic Honesty NAME (Print) Wenge Xu STUDENT NUMBER 201324739 SCHOOL/INSTITUTE University of Liverpool TITLE OF WORK Motion-based Interaction for Head-Mounted Displays This form should be completed by the student and appended to any piece of work that is submitted for examination. Submission by the student of the form by electronic means constitutes their confirmation of the terms of the declaration. Students should familiarise themselves with Appendix 4 of the PGR Code of Practice: PGR Policy on Plagiarism and Dishonest Use of Data, which provides the definitions of academic malpractice and the policies and procedures that apply to the investigation of alleged incidents. Students found to have committed academic malpractice will receive penalties in accordance with the Policy, which in the most severe cases might include termination of studies. STUDENT DECLARATION I confirm that: • I have read and understood the University’s PGR Policy on Plagiarism and Dishonest Use of Data. • I have acted honestly, ethically and professionally in conduct leading to assessment for the programme of study. • I have not copied material from another source nor committed plagiarism nor fabricated, falsified or embellished data when completing the attached material. • I have not copied material from another source, nor colluded with any other student in the preparation and production of this material. • If an allegation of suspected academic malpractice is made, I give permission to the University to use source-matching software to ensure that the submitted material is all my own work.
    [Show full text]
  • Augmented Government Transforming Government Services Through Augmented Reality
    Augmented government Transforming government services through augmented reality A GovLab Study About GovLab GovLab is a think tank in the Federal practice of Deloitte Consulting LLP that focuses on innovation in the public sector. It works closely with senior government executives and thought leaders from across the globe. GovLab Fellows conduct research into key issues and emerging ideas shaping the public, private, and non-profit sectors. Through exploration and analysis of government’s most pressing challenges, GovLab seeks to develop innovative yet practical ways that governments can transform the way they deliver their services and prepare for the challenges ahead. About Deloitte Deloitte refers to one or more of Deloitte Touche Tohmatsu Limited, a UK private company limited by guarantee, and its network of member firms, each of which is a legally separate and independent entity. Please see www. deloitte.com/about for a detailed description of the legal structure of Deloitte Touche Tohmatsu Limited and its member firms. Please see www.deloitte. com/us/about for a detailed description of the legal structure of Deloitte LLP and its subsidiaries. Certain services may not be available to attest clients under the rules and regulations of public accounting. As used in this document, “Deloitte” means Deloitte Consulting LLP, a subsid- iary of Deloitte LLP. Please see www.deloitte.com/us/about for a detailed description of the legal structure of Deloitte LLP and its subsidiaries. ertain services may not be available to attest clients under the rules and regulations of public accounting. Augmented Government Transforming government services through augmented reality 3 Contents 1 Introduction 9 Seeing beyond borders 15 Expanding perimeters 21 The eye of the storm 26 What’s next? 28 Recent public sector thought leadership 29 Endnotes 32 Contacts 4 Introduction On September 27, 1998, football fans who tuned in to watch the game between the Baltimore Ravens and the Cincinnati Bengals witnessed “These new technologies are changing … much more than a 31–24 Bengal win.
    [Show full text]
  • A User Study to Analyse the Experience of Augmented Reality Board Games
    Bachelor’s Thesis in Digital Game Development June 2019 A user study to analyse the experience of augmented reality board games Giedre Jursenaite Daniel Bengtsson Faculty of Computing, Blekinge Institute of Technology, 371 79 Karlskrona, Sweden This thesis is submitted to the Faculty of Computing at Blekinge Institute of Technology in partial fulfillment of the requirements for the degree of Bachelor’s Thesis in Digital Game Development. The thesis is equivalent to 10 weeks of full-time studies. The authors declare that they are the sole authors of this thesis and that they have not used any sources other than those listed in the bibliography and identified as references. They further declare that they have not submitted this thesis at any other institution to obtain a degree. Contact Information: Author(s): Giedre Jursenaite E-mail: [email protected] Daniel Bengtsson E-mail: [email protected] University advisor: M.Sc. Diego Navarro Department of Computer Science Faculty of Computing Internet : www.bth.se Blekinge Institute of Technology Phone : +46 455 38 50 00 SE–371 79 Karlskrona, Sweden Fax : +46 455 38 50 57 Abstract Background. Augmented Reality (AR) is a variant of virtual reality (VR), but where VR replaces reality with a virtual one, AR expands it allowing the user to see the real world and virtual information at the same time. Many have tried to adapt this technology for video and board games and although there are plenty of AR video games or mobile game applications there are no one selling AR board games. Some studies keep coming up from time to time trying to enhance board games with AR through graphics and extra information about player statistics, but there are not many that adapt game logic in AR games.
    [Show full text]
  • Smart Glasses: Interaction, Privacy and Social Implications
    Smart glasses: Interaction, privacy and social implications Ubiquitous Computing Seminar FS2014 Student report Marica Bertarini ETH Zurich Department of Computer Science Zurich, Switzerland [email protected] ABSTRACT INTERACTION Smart glasses are wearable devices that display real-time The main purpose of smart glasses is to provide users with information directly in front of users’ field of vision by us- information and services relevant for their contexts and ing Augmented Reality (AR) techniques. Generally, they useful for the users to perform their tasks; in other words, can also perform more complex tasks, run some applica- such devices augment users’ senses. In addition, they allow tions, and support Internet connectivity. This paper pro- users to do basic operations available on today common vides an overview of some methods that can be adopted to mobile devices such as reading, writing e-mails, writing allow gesture-based interaction with smart glasses, as well text messages, making notes, and answering calls. as of some interaction design considerations. Additionally, Therefore, although most of the usage of smart glasses is it discusses some social effects induced by a wide-spread passive for the users, i.e. reading content on the little screen deployment of smart glasses as well as possible privacy of the device, active interaction with such devices is fun- concerns. damental to control them and supply inputs. In fact, users Keywords: Smart glasses, Head-worn displays, input tech- need ways to ask smart glasses for instance to open a par- niques, interaction, body interaction, in-air interaction, pri- ticular application, answer something they need to know, vacy, social implications insert content for emails, messages or input fields, or to control games.
    [Show full text]
  • Trump Camp Posts Fake Video of Biden () Atlas | Partners in Parkour
    Serving robot finds work at Victoria restaurant () Utopia VR - The Metaverse for Everyone () Utopia VR Promo Video () Facebook warned over ‘very small’ indicator LED on smart glasses, as EU DPAs flag privacy concerns () Google’s Former AI Ethics Chief Has a Plan to Rethink Big Tech () Introducing iPhone 13 Pro | Apple () Introducing Apple Watch Series 7 | Apple () Apple reveals Apple Watch Series 7, featuring a larger, more advanced display () Xiaomi shows off concept smart glasses with MicroLED display () Xiaomi Smart Glasses | Showcase | A display in front of your eyes () ‘Neurograins’ Could be the Next Brain-Computer Interfaces () Welcome back to the moment. With Ray-Ban x Facebook () Facebook's Ray-Ban Stories smart glasses: Cool or creepy? () Introducing WHOOP 4.0 - The Personalized Digital Fitness and Health Coach () Introducing WHOOP Body - The Future of Wearable Technology () The Future of Communication in the Metaverse () Making pancakes with Reachy through VR Teleoperation () Trump Camp Posts Fake Video Of Biden () More than 50 robots are working at Singapore's high-tech hospital () The manual for Facebook’s Project Aria AR glasses shows what it’s like to wear them () Tesla Bot explained () Page 1 of 212 Tesla Bot Takes Tech Demos to Their Logical, Absurd Conclusion () Watch brain implant help man 'speak' for the first time in 15 years () An Artificial Intelligence Helped Write This Play. It May Contain Racism () Elon Musk REVEALS Tesla Bot () Meet Grace, a humanoid robot designed for healthcare () Horizon Workrooms - Remote Collaboration Reimagined () Atlas | Partners in Parkour () Inside the lab: How does Atlas work? () Neural recording and stimulation using wireless networks of microimplants () The Metaverse is a Dystopian Nightmare.
    [Show full text]