Voice Interfaces

Total Page:16

File Type:pdf, Size:1020Kb

Voice Interfaces VIEW POINT VOICE INTERFACES Abstract A voice-user interface (VUI) makes human interaction with computers possible through a voice/speech platform in order to initiate an automated service or process. This Point of View explores the reasons behind the rise of voice interface, key challenges enterprises face in voice interface adoption and the solution to these. Are We Ready for Voice Interfaces? Let’s get talking! IO showed the new promise of voice offering integrations with their Voice interfaces. Assistants. Since Apple integration with Siri, voice interfaces has significantly Almost all the big players (Google, Apple, As per industry forecasts, over the next progressed. Echo and Google Home Microsoft) have ‘office productivity’ decade, 8 out of every 10 people in the have demonstrated that we do not need applications that are being adopted by world will own a device (a smartphone or a user interface to talk to computers businesses (Microsoft and their Office some kind of assistant) which will support and have opened-up a new channel for Suite already have a big advantage here, voice based conversations in native communication. Recent demos of voice but things like Google Docs and Keynote language. Just imagine the impact! based personal assistance at Google are sneaking in), they have also started Voice Assistant Market USD~7.8 Billion CAGR ~39% Market Size (USD Billion) 2016 2017 2018 2019 2020 2021 2022 2023 The Sudden Interest in Voice Interfaces Although voice technology/assistants Voice Recognition Accuracy Convenience – speaking vs typing have been around in some shape or form Voice Recognition accuracy continues to Humans can speak 150 words per minute for many years, the relentless growth of improve as we now have the capability to vs the typing speed of 40 words per low-cost computational power—and train the models using neural networks minute. This makes voice interfaces an breakthroughs in machine learning and and large amount of relevant user data. efficient way to interact with technology. deep learning—mean bots can now be Google has been able to achieve 95% With natural language support and ability built with self-learning capabilities. machine learning word accuracy which to respond intelligently to user queries is the same as human accuracy. It is also based upon user’s context, behavior expected to receive 50% of all searches in patterns and AI based learnings, voice Big breakthroughs in AI/ML voice by 2020 which will further improve based systems are increasingly becoming space in the recent years have recognition across languages as it will give an exciting option for users as is evident improved the voice recognition access to variety of new data for training. from the ever increasing sales numbers accuracy, adoption has also A lot of recent advancements in the NLP of voice assistants (Amazon Echo, Google increased as it is convenient (Natural Language Processing), NLG Home etc.). to use voice interfaces, as a (Natural Language Generation) and Text result a lot of interest and to Speech(TTS) areas have also given a big investments are attracted. push to voice based applications. External Document © 2019 Infosys Limited External Document © 2019 Infosys Limited Industry investments Google Machine Learning Word Accuracy Tech giants have entered the business of Voice Interfaces through either direct 100% investments or acquisitions. For example, Google recently acquired Mobvoi (founded 95% 95% by Li Zhifei, a former Google Research 90% Scientist) while Apple acquired Siri. Rate Amazon’s Bezos has openly said that they cy have more than 1,000 people working on Alexa and the Echo ecosystem for the past rd Accura 80% many years. Bezos, through his investment Wo fund Bezos Expeditions, and the Alexa Fund have invested in a $2.5 million seed round for Pulse Labs led by the Seattle- 70% 2013 2014 2015 2016 2017 based venture firm Madrona Venture Group. Google Threshold for Human Accuracy Key Technologies Enabling Voice Interfaces Currently there are four key enablers for voice interfaces: Automatic Speech Natural Language Natural Language Text to Speech (TTS) Recognition (ASR) Understanding Generation Each of these technologies have seen big options. Enterprises can also embrace business/use case requirements breakthroughs in recent years providing either online (cloud services) or offline enterprises with a multitude of solutions/ (on-premise) option depending on their Natural Language Understanding Intent Entity Recognition Extraction Speech Recognition Conversation Session and Management Context Management Response Generation Fullment Streaming Speech Natural Output Text to Speech Language Generation External Document © 2019 Infosys Limited External Document © 2019 Infosys Limited Key players and options for enterprises The vendor landscape for Voice Interfaces on-premise, the issue of routing users • While open-source on-premise options has become quite crowded with the IT and business data via third-party/cloud for speech recognition are evolving/ giants of the world having the majority based services is taken care of improving fast, there are commercial share of the pie. Here is a quick look. options available like Nuance, NICE etc. • Licensing cost, cost of maintaining/ On Premise Deployment scaling the infrastructure will vary depending upon the nature of use case/ • When speech recognition is brought volume of transactions On-premise Deployment Corporate Boundary Congure and Train Language Model Bot Conguration Designer/Developer Refer Conguration and Trained Model API Creates and Deploys Process Intent Bot r ye Enterprise Application La tegration Systems Enterprise In Recognize Intent and Extract Entities User Speaks Enterprise Speech Apps Recognition User Cloud Deployment • There are advantages of pay as you powered Google Speech API for Speech use, ease of on demand scale, no to text conversion that is available for • Cloud service providers offer a promising infrastructure/platform maintenance short or long-form audio with high option by allowing an enterprise to overhead, inbuilt analytics etc. accuracy recognition, Amazon has configure and deploy voice bots on Transcribe and Microsoft comes with • Leading online speech service providers cloud (server-less as lambda/cloud Microsoft Cognitive Services and Bing are the giants like Google, Amazon and functions or server infrastructure) which Speech API. can integrate with the enterprise systems Microsoft. While Google provides ML Cloud Deployment Corporate Boundary Cloud Congure and Train Language Model Designer/Developer Bot Conguration API User Speaks ices Channels Process rv Intent Se r User NLP/NLU ye Enterprise La tegration Speech Recognition Systems Enterprise Recognize Intent and In Extract Entities Channels/Cloud Enterprise Apps Service Providers User Speaks External Document © 2019 Infosys Limited External Document © 2019 Infosys Limited The following diagram shows the various players in on-premise, cloud and data set services Speech Recognition: A view of Online, Open Source and Dataset options Google Cloud Speech Mozilla Deep Speech API RNN Based Model Amazon Transcribe Facebook Wave2letter CNN Based Model vices Microsoft Bing Speech Open Source API Kaldi DNN Based Model Cloud Ser IBM Watson Speech API Nuance Datasets LibriSpeech SwitchBoard Google Command Fisher Enterprise Readiness for Voice as Their Strategic Priority Redesigning/enhancing digital experience organizations while conversational AI with key for businesses to enhance customers is by far the top strategic priority for most voice recognition capability is becoming experience and attract new customers. Challenges Speech Recognition owing to its that separates their world-class speech no cost but most importantly they are complexity has taken years of effort to recognition system from any other speech getting real life users’ audio data for further evolve to this level of maturity. There recognition system created elsewhere. training to improve their experience. This are multiple challenges with users’ real Alexa, Google Home, Apple Home Pod, Siri, drives enterprises to depend on such cloud life audio streams including low quality Google Now are providing convenience players in developing voice interfaces for microphones, background noise, reverb to the customer at minimal one-time or enterprise applications. and echo, accent variations etc. Training data should contain samples with all Speech Recognition these problems in order to make sure the neural network can deal with them. Other Neural Network challenges that enterprises face include: Data set for training Model Training Accuracy of machine learning solutions Input Data is completely driven by the amount of available data sets for training. As for the voice recognition accuracy, it is difficult for enterprises to create their own approach and solution because of their inability to User says gather variety and volume of data required “Hello” to build the right models. Recognition IT giants like Google, Amazon or Baidu Text Response do possess huge datasets with hours and hours of audio recordings in real-life “Hello” situations. It is the single biggest factor External Document © 2019 Infosys Limited External Document © 2019 Infosys Limited Lack of mature open source have the potential to reach a good level users’ data and other critical business options of maturity. Mozilla has launched Project information) being routed via cloud Common Voice, an initiative to help make providers for voice recognition. Some While
Recommended publications
  • The Chatbot Revolution
    The chatbot revolution Moving beyond the hype and maximizing customer experience Ready for a digital 2% 2017 conversation? 25% 2020 Consumers have higher expectations than ever when it comes to interacting with brands. By 2020, 25% of customer They demand personalized service, easily accessible support options, a quick response after reaching out, and successful resolutions on a tight turnaround. service operations will use To meet these needs, companies are increasing their use of digital channels to chatbot or virtual assistant communicate with customers – in fact, by 2022, 70% of all customer interactions will involve technology like messaging applications, social platforms, or chatbots. technologies, an increase Let’s take a closer look at chatbots. Their functions range from answering simple from just 2% in 2017. questions like informing customers of store hours or location to more advanced ones, like handling a credit card charge dispute. According to Gartner, by 2020, 25% of customer service operations will use chatbot or virtual assistant technologies, an increase from just 2% in 2017. When trying to balance staffing budgets, round- the-clock service availability and a preference for digital platforms, chatbots on paper seem like the obvious – and inevitable – choice to engage customers through automation. But how inevitable is it really? 1. Gartner Magic Quadrant for Customer Engagement Center, Michael Maoz, Brian Manusama, 16 May 2018 www.pega.com The chatbot revolution 01 Why the digital hold up? Consumers and businesses have concerns. Despite Gartner predictions and the obvious excitement around chatbots, overall adoption has been slow. Currently most chatbots are programmed to follow predetermined conversational flows—thus limiting their usefulness for solving complex problems or picking up conversational subtleties.
    [Show full text]
  • Intellibot: a Domain-Specific Chatbot for the Insurance Industry
    IntelliBot: A Domain-specific Chatbot for the Insurance Industry MOHAMMAD NURUZZAMAN A thesis submitted in fulfilment of the requirements for the degree of Doctor of Philosophy UNSW Canberra at Australia Defence Force Academy (ADFA) School of Business 20 October 2020 ORIGINALITY STATEMENT ‘I hereby declare that this submission is my own work and to the best of my knowledge it contains no materials previously published or written by another person, or substantial proportions of material which have been accepted for the award of any other degree or diploma at UNSW or any other educational institute, except where due acknowledgement is made in the thesis. Any contribution made to the research by others, with whom I have worked at UNSW or elsewhere, is explicitly acknowledged in the thesis. I also declare that the intellectual content of this thesis is the product of my own work, except to the extent that assistance from others in the project’s design and conception or in style, presentation and linguistic expression is acknowledged.’ Signed Date To my beloved parents Acknowledgement Writing a thesis is a great process to review not only my academic work but also the journey I took as a PhD student. I have spent four lovely years at UNSW Canberra in the Australian Defence Force Academy (ADFA). Throughout my journey in graduate school, I have been fortunate to come across so many brilliant researchers and genuine friends. It is the people who I met shaped who I am today. This thesis would not have been possible without them. My gratitude goes out to all of them.
    [Show full text]
  • The Inner Circle Guide to AI, Chatbots & Machine Learning
    The Inner Circle Guide to AI, Chatbots & Machine Learning Sponsored by The Inner Circle Guide to AI, Chatbots and Machine Learning © ContactBabel 2019 Please note that all information is believed correct at the time of publication, but ContactBabel does not accept responsibility for any action arising from errors or omissions within the report, links to external websites or other third-party content. 2 Understand the customer experience with the power of AI Employees Customers Businesses Increase agent Elevate customer Gain improved engagement experience VoC insights Artificial Customer Machine Intelligence surveys learning Recorded CRM VoC calls notes analytics Social media Chatbots Surveys opentext.com/explore CONTENTS Contents ..................................................................................................................................................... 4 Table of Figures ........................................................................................................................................... 6 About the Inner Circle Guides ..................................................................................................................... 7 AI: Definitions and Terminology ................................................................................................................. 9 Definitions............................................................................................................................................. 11 Use Cases for AI in the Contact Centre ....................................................................................................
    [Show full text]
  • NLP-5X Product Brief
    NLP-5x Natural Language Processor With Motor, Sensor and Display Control The NLP-5x is Sensory’s new Natural Language Processor targeting consumer electronics. The NLP-5x is designed to offer advanced speech recognition, text-to-speech (TTS), high quality music and speech synthesis and robotic control to cost-sensitive high volume products. Based on a 16-bit DSP, the NLP-5x integrates digital and analog processing blocks and a wide variety of communication interfaces into a single chip solution minimizing the need for external components. The NLP-5x operates in tandem with FluentChip™ firmware - an ultra- compact suite of technologies that enables products with up to 750 seconds of compressed speech, natural language interface grammars, TTS synthesis, Truly Hands-Free™ triggers, multiple speaker dependent and independent vocabularies, high quality stereo music, speaker verification (voice password), robotics firmware and all application code built into the NLP-5x as a single chip solution. The NLP-5x also represents unprecedented flexibility in application hardware designs. Thanks to the highly integrated architecture, the most cost-effective voice user interface (VUI) designs can be built with as few additional parts as a clock crystal, speaker, microphone, and few resistors and capacitors. The same integration provides all the necessary control functions on-chip to enable cost-effective man-machine interfaces (MMI) with sensing technologies, and complex robotic products with motors, displays and interactive intelligence. Features BROAD
    [Show full text]
  • A Survey on Different Algorithms Used in Chatbot
    International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 07 Issue: 05 | May 2020 www.irjet.net p-ISSN: 2395-0072 A survey on Different Algorithms used in Chatbot Siddhi Pardeshi1, Suyasha Ovhal2, Pranali Shinde3, Manasi Bansode4, Anandkumar Birajdar5 1Siddhi Pardeshi, Student, Dept. of Computer Engineering, Pimpri Chinchwad College of Engineering, Pune Maharashtra, India 2Suyasha Ovhal, Student, Dept. of Computer Engineering, Pimpri Chinchwad College of Engineering, Pune Maharashtra, India 3Pranali Shinde, Student, Dept. of Computer Engineering, Pimpri Chinchwad College of Engineering, Pune Maharashtra, India 4Manasi Bansode, Student, Dept. of Computer Engineering, Pimpri Chinchwad College of Engineering, Pune Maharashtra, India 5Professor, Dept. of Computer Engineering, Pimpri Chinchwad College of Engineering, Pune, Maharashtra, India ---------------------------------------------------------------------***---------------------------------------------------------------------- Abstract - Machines are working similar to humans Rule-based/Command based: In these types of chatbots, because of advanced technological concepts. Best example is predefined rules are stored which includes questions and chatbot which depends on advanced concepts in computer answers. Based on what question has requested by the user science. Chatbots serve as a medium for the communication chatbot searches for an answer. But this gives limitations on between human and machine. There are a number of chatbots the type of questions and answers to be stored. and design techniques available in market that perform Intelligent Chatbots/AI Chatbots: To overcome the issue different function and can be implemented in sectors like faced by rule based chatbots intelligent chatbots are business sector, medical sector, farming etc. The technology developed. As these are based on advanced machine learning used for the advancement of conversational agent is natural concepts, they have the ability to learn on their own language processing (NLP).
    [Show full text]
  • A Guide to Chatbot Terminology
    Machine Language A GUIDE TO CHATBOT TERMINOLOGY Decision Trees The most basic chatbots are based on tree-structured flowcharts. Their responses follow IF/THEN scripts that are linked to keywords and buttons. Natural Language Processing (NLP) A computer’s ability to detect human speech, recognize patterns in conversation, and turn text into speech. Natural Language Understanding A computer’s ability to determine intent, especially when what is said doesn’t quite match what is meant. This task is much more dicult for computers than Natural Language Processing. Layered Communication Human communication is complex. Consider the intricacies of: • Misused phrases • Intonation • Double meanings • Passive aggression • Poor pronunciation • Regional dialects • Subtle humor • Speech impairments • Non-native speakers • Slang • Syntax Messenger Chatbots Messenger chatbots reside within the messaging applications of larger digital platforms (e.g., Facebook, WhatsApp, Twitter, etc.) and allow businesses to interact with customers on the channels where they spend the most time. Chatbot Design Programs There’s no reason to design a messenger bot from scratch. Chatbot design programs help designers make bots that: • Can be used on multiple channels • (social, web, apps) • Have custom design elements • (response time, contact buttons, • images, audio, etc.) • Collect payments • Track analytics (open rates, user • retention, subscribe/unsubscribe • rates) • Allow for human takeover when • the bot’s capabilities are • surpassed • Integrate with popular digital • platforms (Shopify, Zapier, Google • Site Search, etc.) • Provide customer support when • issues arise Voice User Interface A voice user interface (VUI) allows people to interact with a computer through spoken commands and questions. Conversational User Interface Like a voice user interface, a conversational user interface (CUI) allows people to control a computer with speech, but CUI’s dier in that they emulate the nuances of human conversation.
    [Show full text]
  • MULTILINGUAL CHATBOT with HUMAN CONVERSATIONAL ABILITY [1] Aradhana Bisht, [2] Gopan Doshi, [3] Bhavna Arora, [4] Suvarna Pansambal [1][2] Student, Dept
    International Journal of Future Generation Communication and Networking Vol. 13, No. 1s, (2020), pp. 138- 146 MULTILINGUAL CHATBOT WITH HUMAN CONVERSATIONAL ABILITY [1] Aradhana Bisht, [2] Gopan Doshi, [3] Bhavna Arora, [4] Suvarna Pansambal [1][2] Student, Dept. of Computer Engineering,[3][4] Asst. Prof., Dept. of Computer Engineering, Atharva College of Engineering, Mumbai, India Abstract Chatbots - The chatbot technology has become very fascinating to people around the globe because of its ability to communicate with humans. They respond to the user query and are sometimes capable of executing sundry tasks. Its implementation is easier because of wide availability of development platforms and language libraries. Most of the chatbots support English language only and very few have the skill to communicate in multiple languages. In this paper we are proposing an idea to build a chatbot that can communicate in as many languages as google translator supports and also the chatbot will be capable of doing humanly conversation. This can be done by using various technologies such as Natural Language Processing (NLP) techniques, Sequence To Sequence Modeling with encoder decoder architecture[12]. We aim to build a chatbot which will be like virtual assistant and will have the ability to have conversations more like human to human rather than human to bot and will also be able to communicate in multiple languages. Keywords: Chatbot, Multilingual, Communication, Human Conversational, Virtual agent, NLP, GNMT. 1. Introduction A chatbot is a virtual agent for conversation, which is capable of answering user queries in the form of text or speech. In other words, a chatbot is a software application/program that can chat with a user on any topic[5].
    [Show full text]
  • A Chatbot System Demonstrating Intelligent Behaviour Using
    International Journal of Advanced Research in Computer Engineering & Technology (IJARCET) Volume 4 Issue 10, October 2015 A chatbot system demonstrating Intelligent Behaviour using NLP 1Ameya Vichare, 2Ankur Gyani, 3Yashika Shrikhande, 4Nilesh Rathod 1Student, IT Department, RGIT, Mumbai 2Student, IT Department, RGIT, Mumbai 3Student, IT Department, RGIT, Mumbai 4Assistant Professor, IT Department, RGIT, Mumbai Abstract— Chatbots are computer programs that interact making capabilities, availability of corpora, processing tool with users using natural languages. Just as people use language standards like XML [1]. for human communication, chatbots use natural language to communicate with human users. In this paper, we begin by In our project we are using AIML. AIML is an XML- introducing chatbots and emphasize their need in day-to-day based language which can be used for interaction between activities. Then we go on to discuss existing chatbot systems, chatbots and humans. The atomic unit in AIML is category, namely ELIZA, ALICE and Siri. We evaluate what we can take which consists of attributes called as pattern and template. from each of the systems and use in our proposed system. We are also using a speech to text/text to speech recognition Finally, we discuss the proposed system. The system we intend to develop can be used for educating the user about sports. The to recognize the Indian accent more efficiently. While database will be fed with sports related data which will be constructing the soundbase of the chatbot, the following can coded using AIML. This system can be used in operating help heighten its speech recognition rate: the soundbase systems in a similar manner to Siri for effective information should be built to match user speech input based on retrieval just by speaking various queries.
    [Show full text]
  • Voice User Interface on the Web Human Computer Interaction Fulvio Corno, Luigi De Russis Academic Year 2019/2020 How to Create a VUI on the Web?
    Voice User Interface On The Web Human Computer Interaction Fulvio Corno, Luigi De Russis Academic Year 2019/2020 How to create a VUI on the Web? § Three (main) steps, typically: o Speech Recognition o Text manipulation (e.g., Natural Language Processing) o Speech Synthesis § We are going to start from a simple application to reach a quite complex scenario o by using HTML5, JS, and PHP § Reminder: we are interested in creating an interactive prototype, at the end 2 Human Computer Interaction Weather Web App A VUI for "chatting" about the weather Base implementation at https://github.com/polito-hci-2019/vui-example 3 Human Computer Interaction Speech Recognition and Synthesis § Web Speech API o currently a draft, experimental, unofficial HTML5 API (!) o https://wicg.github.io/speech-api/ § Covers both speech recognition and synthesis o different degrees of support by browsers 4 Human Computer Interaction Web Speech API: Speech Recognition § Accessed via the SpeechRecognition interface o provides the ability to recogniZe voice from an audio input o normally via the device's default speech recognition service § Generally, the interface's constructor is used to create a new SpeechRecognition object § The SpeechGrammar interface can be used to represent a particular set of grammar that your app should recogniZe o Grammar is defined using JSpeech Grammar Format (JSGF) 5 Human Computer Interaction Speech Recognition: A Minimal Example const recognition = new window.SpeechRecognition(); recognition.onresult = (event) => { const speechToText = event.results[0][0].transcript;
    [Show full text]
  • Eindversie-Paper-Rianne-Nieland-2057069
    Talking to Linked Data: Comparing voice interfaces for general­purpose data Master thesis of Information Sciences Rianne Nieland Vrije Universiteit Amsterdam [email protected] ABSTRACT ternet access (Miniwatts Marketing Group, 2012) and People in developing countries cannot access informa- 31.1% is literate (UNESCO, 2010). tion on the Web, because they have no Internet access and are often low literate. A solution could be to pro- A solution to the literacy and Internet access problem vide voice-based access to data on the Web by using the is to provide voice-based access to the Internet by us- GSM network. Related work states that Linked Data ing the GSM network (De Boer et al., 2013). 2G mobile could be a useful input source for such voice interfaces. phones are digital mobile phones that use the GSM net- work (Fendelman, 2014). In Africa the primary mode The aim of this paper is to find an efficient way to make of telecommunication is mobile telephony (UNCTAD, general-purpose data, like Wikipedia information, avail- 2007). able using voice interfaces for GSM. To achieve this, we developed two voice interfaces, one for Wikipedia This paper is about how information on the Web ef- and one for DBpedia, by doing requirements elicitation ficiently can be made available using voice interfaces from literature and developing a voice user interface and for GSM. We developed two voice interfaces, one us- conversion algorithms for Wikipedia and DBpedia con- ing Wikipedia and the other using DBpedia. In this cepts. With user tests the users evaluated the two voice paper we see Wikipedia and DBpedia as two different interfaces, to be able to compare them.
    [Show full text]
  • Voice Assistants and Smart Speakers in Everyday Life and in Education
    Informatics in Education, 2020, Vol. 19, No. 3, 473–490 473 © 2020 Vilnius University, ETH Zürich DOI: 10.15388/infedu.2020.21 Voice Assistants and Smart Speakers in Everyday Life and in Education George TERZOPOULOS, Maya SATRATZEMI Department of Applied Informatics, University of Macedonia, Thessaloniki, Greece Email: [email protected], [email protected] Received: November 2019 Abstract. In recent years, Artificial Intelligence (AI) has shown significant progress and its -po tential is growing. An application area of AI is Natural Language Processing (NLP). Voice as- sistants incorporate AI by using cloud computing and can communicate with the users in natural language. Voice assistants are easy to use and thus there are millions of devices that incorporates them in households nowadays. Most common devices with voice assistants are smart speakers and they have just started to be used in schools and universities. The purpose of this paper is to study how voice assistants and smart speakers are used in everyday life and whether there is potential in order for them to be used for educational purposes. Keywords: artificial intelligence, smart speakers, voice assistants, education. 1. Introduction Emerging technologies like virtual reality, augmented reality and voice interaction are reshaping the way people engage with the world and transforming digital experiences. Voice control is the next evolution of human-machine interaction, thanks to advances in cloud computing, Artificial Intelligence (AI) and the Internet of Things (IoT). In the last years, the heavy use of smartphones led to the appearance of voice assistants such as Apple’s Siri, Google’s Assistant, Microsoft’s Cortana and Amazon’s Alexa.
    [Show full text]
  • Integrating a Voice User Interface Into a Virtual Therapy Platform Yun Liu Lu Wang William R
    Integrating a Voice User Interface into a Virtual Therapy Platform Yun Liu Lu Wang William R. Kearns University of Washington, Seattle, University of Washington, Seattle, University of Washington, Seattle, WA 98195, USA WA 98195, USA WA 98195, USA [email protected] [email protected] [email protected] Linda E Wagner John Raiti Yuntao Wang University of Washington, Seattle, University of Washington, Seattle, Tsinghua University, Beijing 100084, WA 98195, USA WA 98195, USA China & University of Washington, [email protected] [email protected] WA 98195, USA [email protected] Weichao Yuwen University of Washington, Tacoma, WA 98402, USA [email protected] ABSTRACT ACM Reference Format: More than 1 in 5 adults in the U.S. serving as family caregivers are Yun Liu, Lu Wang, William R. Kearns, Linda E Wagner, John Raiti, Yuntao Wang, and Weichao Yuwen. 2021. Integrating a Voice User Interface into a the backbones of the healthcare system. Caregiving activities sig- Virtual Therapy Platform. In CHI Conference on Human Factors in Computing nifcantly afect their physical and mental health, sleep, work, and Systems Extended Abstracts (CHI ’21 Extended Abstracts), May 08–13, 2021, family relationships over extended periods. Many caregivers tend Yokohama, Japan. ACM, New York, NY, USA, 6 pages. https://doi.org/10. to downplay their own health needs and have difculty accessing 1145/3411763.3451595 support. Failure to maintain their own health leads to diminished ability in providing high-quality care to their loved ones. Voice user interfaces (VUIs) hold promise in providing tailored support 1 INTRODUCTION family caregivers need in maintaining their own health, such as Chronic diseases, such as diabetes and cancer, are the leading causes fexible access and handsfree interactions.
    [Show full text]