Comparing Voice and Touch Interaction for Smartphone Radio and Podcast Application

Total Page:16

File Type:pdf, Size:1020Kb

Comparing Voice and Touch Interaction for Smartphone Radio and Podcast Application DEGREE PROJECT IN THE FIELD OF TECHNOLOGY MEDIA TECHNOLOGY AND THE MAIN FIELD OF STUDY COMPUTER SCIENCE AND ENGINEERING, SECOND CYCLE, 30 CREDITS STOCKHOLM, SWEDEN 2017 Comparing voice and touch interaction for smartphone radio and podcast application FREDRIK WALLÉN KTH ROYAL INSTITUTE OF TECHNOLOGY SCHOOL OF COMPUTER SCIENCE AND COMMUNICATION Comparing voice and touch interaction for smartphone radio and podcast application Jämförelse av röst- och pekskärmsinteraktion för en radio- och podcastapplikation för smartphones Fredrik Wallén KTH Royal Institute of Technology Stockholm, Sweden [email protected] Subject: Media Technology Master of Science (in engineering) in Media Technology Master in Media Technology Supervisor: Filip Kis Examiner: Roberto Bresin Principal: Isotop 29 June 2017 SAMMANFATTNING Röststyrningen blir vanligare och numera är den också möjligt att använda i individuella appar för smartphones. Det har dock inte tidigare undersökts för vilka uppgifter det ur ett användbarhetsperspektiv är att föredra framför pekskärmsinteraktion. För att undersöka det skapades ett röstinterface för en radio och podcast applikation som redan hade ett pekskärmsinterface. Röstinterfacet testades också med användare för att förbättra dess användbarhet. Efter det gjordes ett test där deltagarna blev ombedda att utföra samma uppgift med både pekskärm- och röstinterface. Den tid de tog på sig uppmättes och deltagarna betygsatte upplevelsen av att utföra uppgiften på en skala. Slutligen blev de tillfrågade om vilken interaktionsmetod de föredrog. För de flesta av de testade uppgifterna var röstinteraktion snabbare och fick högre betyg. Det ska dock noteras att i fall då användaren inte har specifika uppgifter att utföra kan det vara svårare för dem att veta vad en röststyrd app kan och inte kan göra än när de använder pekskärm. Många användare uttryckte också att de var motvilliga till att använda röstkommandon i allmänna utrymmen av rädsla för att verka underliga. Dessa resultat kan tillämpas på radio/podcast appar och, i mindre utsträckning, appar för att titta på TV-serier och spela musik. Comparing voice and touch interaction for smartphone radio and podcast application Fredrik Wallén KTH Royal Institute of Technology Stockholm, Sweden [email protected] ABSTRACT of NLUIs and VUIs in which mobile apps like Siri, Google Now Today voice recognition is becoming mainstream and nowadays it and Cortana, which combine visual and auditory information and is also possible to include in individual smartphone apps. voice-only devices such as the Amazon Echo and Google Home However, it has not previously been investigated for which tasks are becoming mainstream. it is preferable from a usability perspective to use voice Voice control is no longer exclusive to dedicated voice control recognition rather than touch. In order to investigate this, a voice apps. It is now possible for developers to use a voice control user interface was created for a smartphone radio application, interface for any app. However, it has not previously been which already had a touch interface. The voice user interface was investigated in which cases it is beneficial to do so from a also tested with users in order to improve its usability. After that, usability perspective. a test was conducted where the participants were asked to perform the same tasks using both the touch and voice interface. The time 1.2 Problem definition they took to complete the tasks was measured and the participants The objective of the study is to investigate for which types of rated the experience of completing the task on a scale. Finally, tasks, within a smartphone radio application, using voice they were asked which interaction method they preferred. commands provides a better user experience than using the For most of the tasks tested, the voice interaction was both faster touchscreen. and got a higher rating. However, it should be noted that in a case 1.3 Expected scientific results where users don’t have specific tasks to perform it might be My hypothesis was that touch interaction would be faster and give harder for them to know what a voice controlled app can and a better user experience than touch interaction except in cases cannot do than when they are using touch. Many users also where touch interaction would require several taps and keyboard expressed that they were reluctant to use voice commands in input. In these cases, voice interaction would be faster and and public spaces out of fear of appearing strange. These results can give a better user experience. be applied to other radio/podcast apps and, to a lesser extent, app for watching TV series and playing music. 1.4 Principal The thesis was carried out at Isotop, an IT consultancy company Keywords based in Stockholm, Sweden. They are hired by the Swedish voice user interface, voice command, natural language user Radio (SR) to create the Sveriges Radio Play app for iOS and interface, voice search, voice assistant Android. 1. INTRODUCTION 2. THEORY AND CONCEPTS 1.1 Background 2.1 Usability testing - Voice user interfaces At the end of the 1980s so called natural language user interfaces In her book “Designing Voice User Interfaces: Principles of (NLUIs) started to appear, according to Michos et al (1996). They Conversational Experiences” (2016) Cathy Pearl gives advice on provided an alternative to the standard command line interface how to conduct usability testing for voice user interfaces. She used for example in UNIX and DOS. Rather than having to mentions several principles that applies to usability testing in remember that the command for creating a directory in Unix is general, such as testing on users as similar to the target group as “mkdir”, a user could just write: “Create a directory called ​ ​ possible and studying what is good and bad in existing interfaces. Documents” and get the same result. This would take longer to She also advises not to have all users perform the tasks in the write but would make it possible for users not familiar with the same order, since the earlier tasks performed might influence how command language to use the system. people perform later tasks. Instead, it is better to use a latin square According to Pearl (2016) Voice user interfaces (VUIs) became so that the task order is varied. more common in the early 2000s in the form of interactive voice It is a good idea to ask the participants questions both after each response (IVR) systems. These were phone systems in which the task and at the end of the test. Quantitative questions commonly users could talk to a computer and perform the same tasks they use the Likert Scale (“strongly disagree” to “strongly agree”) and would previously perform talking to a human operator over the the questions alternate between being positive (“The system is phone. These systems were often designed to imitate human easy to use”) and negative (“The system is confusing”). conversations and the user commonly replied to questions asked Additionally, there could be open-ended questions such as “How by the computer. do you think the system could be improved?”. Pearl claims that we are in now could be known as the second era 1 Specifically for testing of voice user interfaces, she recommends actions that with certain probabilities will lead to other states. writing the tasks carefully and avoiding mentioning command States are also associated with a reward. You start at a specific words or strategies for completing the tasks. The tasks should also state and can then calculate which actions to take to maximize the not give away too much and only provide the essential probability to get a certain reward. With a partially observable information the user needs to complete them. Additionally, it is Markov decision process, you do not know for certain which state important to test whether the users understand that they can talk to you start at. Instead, there are certain probabilities you start at the system. Apart from this, the common usability testing each state. This makes for a more complicated calculation of how methodology think-aloud, in which the user speaks what they are to get the highest reward. experiencing, doesn’t work very well for testing VUIs since the The rule-based framework, according to Bellegarda, is rooted in users speak to interact with the system. artificial intelligence. It is based on rule sets where each rule In the early stages of design, Dybkjær and Bernsen (2001) consists of a condition and an action. Data and events from user proposes using a “Wizard of Oz” test, which is a test of something input are also inserted into a facts store, which manages facts. For that actually doesn’t work and a human behind the scenes gives example, regarding meetings there might be a rule saying that the illusion of a working system. According to them, this can be they shall contain a date, one or more persons and a location and used to gather valuable data. They emphasize the need to start the that they might contain a topic. evaluation as early as possible and continue evaluating throughout Bellegarda also argue that the statistical approach is better for the development. understanding free speech while the rule based framework is 2.2 Designing voice user interfaces better for understanding input for specific tasks. He also states that Apple’s Siri uses both of these two frameworks and combines Pearl (2016) explains that voice controlled apps have many them to try to get the best outcome. similarities with so called IVR systems which are bots spoken to over the phone. The main difference between Voice Controlled According to Tur and De Mori, in 2011, in commercial apps and IVR systems is that voice controlled apps also have a application the rule-based framework was usually used while the visual component, which can be used to convey additional statistical framework was more commonly used in research.
Recommended publications
  • Voice Interfaces
    VIEW POINT VOICE INTERFACES Abstract A voice-user interface (VUI) makes human interaction with computers possible through a voice/speech platform in order to initiate an automated service or process. This Point of View explores the reasons behind the rise of voice interface, key challenges enterprises face in voice interface adoption and the solution to these. Are We Ready for Voice Interfaces? Let’s get talking! IO showed the new promise of voice offering integrations with their Voice interfaces. Assistants. Since Apple integration with Siri, voice interfaces has significantly Almost all the big players (Google, Apple, As per industry forecasts, over the next progressed. Echo and Google Home Microsoft) have ‘office productivity’ decade, 8 out of every 10 people in the have demonstrated that we do not need applications that are being adopted by world will own a device (a smartphone or a user interface to talk to computers businesses (Microsoft and their Office some kind of assistant) which will support and have opened-up a new channel for Suite already have a big advantage here, voice based conversations in native communication. Recent demos of voice but things like Google Docs and Keynote language. Just imagine the impact! based personal assistance at Google are sneaking in), they have also started Voice Assistant Market USD~7.8 Billion CAGR ~39% Market Size (USD Billion) 2016 2017 2018 2019 2020 2021 2022 2023 The Sudden Interest in Voice Interfaces Although voice technology/assistants Voice Recognition Accuracy Convenience – speaking vs typing have been around in some shape or form Voice Recognition accuracy continues to Humans can speak 150 words per minute for many years, the relentless growth of improve as we now have the capability to vs the typing speed of 40 words per low-cost computational power—and train the models using neural networks minute.
    [Show full text]
  • An Explorative Customer Experience Study on Voice Assistant Services of a Swiss Tourism Destination
    Athens Journal of Tourism - Volume 7, Issue 4, December 2020 – Pages 191-208 Resistance to Customer-driven Business Model Innovations: An Explorative Customer Experience Study on Voice Assistant Services of a Swiss Tourism Destination By Anna Victoria Rozumowski*, Wolfgang Kotowski± & Michael Klaas‡ For tourism, voice search is a promising tool with a considerable impact on tourist experience. For example, voice search might not only simplify the booking process of flights and hotels but also change local search for tourist information. Against this backdrop, our pilot study analyzes the current state of voice search in a Swiss tourism destination so that providers can benefit from those new opportunities. We conducted interviews with nine experts in Swiss tourism marketing. They agree that voice search offers a significant opportunity as a new and diverse channel in tourism. Moreover, this technology provides new marketing measures and a more efficient use of resources. However, possible threats to this innovation are data protection regulation and providers’ lack of skills and financial resources. Furthermore, the diversity of Swiss dialects pushes voice search to its limits. Finally, our study confirms that tourism destinations should cooperate to implement voice search within their touristic regions. In conclusion, following our initial findings from the sample destination, voice search remains of minor importance for tourist marketing in Switzerland as evident in the given low use of resources. Following this initial investigation of voice search in a Swiss tourism destination, we recommended conducting further qualitative interviews on tourists’ voice search experience in different tourist destinations. Keywords: Business model innovation, resistance to innovation, customer experience, tourism marketing, voice search, Swiss destination marketing, destination management Introduction Since innovations like big data and machine learning have already caused lasting changes in the interaction between companies and consumers (Shankar et al.
    [Show full text]
  • The Nao Robot As a Personal Assistant
    1 Ziye Zhou THE NAO ROBOT AS A PERSONAL ASSISTANT Information Technology 2016 2 FOREWORD I would like to take this opportunity to express my gratitude to everyone who have helped me. Dr. Yang Liu is my supervisor in this thesis, without his help I could not come so far and get to know about Artificial Intelligence, I would not understand the gaps between me and other intelligent students in the world. Thanks for giving me a chance to go aboard and get to know more. Also I would like to say thank you to all the teachers and stuffs in VAMK. Thanks for your guidance. Thanks for your patient and unselfish dedication. Finally, thanks to my parents and all my friends. Love you all the time. 3 VAASAN AMMATTIKORKEAKOULU UNIVERSITY OF APPLIED SCIENCES Degree Program in Information Technology ABSTRACT Author Ziye Zhou Title The NAO Robot as a Personal Assistant Year 2016 Language English Pages 55 Name of Supervisor Yang Liu Voice recognition and Artificial Intelligence are popular research topics these days, with robots doing physical and complicated labour work instead of humans is the trend in the world in future. This thesis will combine voice recognition and web crawler, let NAO robot help humans check information (train tickets and weather) from websites by voice interaction with human beings as a voice assistance. The main research works are as follows: 1. Voice recognition, resolved by using Google speech recognition API. 2. Web crawler, resolved by using Selenium to simulate the operation of web pages. 3. The connection and use of NAO robot.
    [Show full text]
  • International Journal of Information Movment
    International Journal of Information Movement Vol.2 Issue III (July 2017) Website: www.ijim.in ISSN: 2456-0553 (online) Pages 85-92 INFORMATION RETRIEVAL AND WEB SEARCH Sapna Department of Library and Information Science Central University of Haryana Email: sapnasna121@gmail Abstract The paper gives an overview of search techniques used for information retrieval on the web. The features of selected search engines and the search techniques available with emphasis on advanced search techniques are discussed. A historic context is provided to illustrate the evolution of search engines in the semantic web era. The methodology used for the study is review of literature related to various aspects of search engines and search techniques available. In this digital era library and information science professionals should be aware of various search tools and techniques available so that they will be able to provide relevant information to users in a timely and effective manner and satisfy the fourth law of library science i.e. “Save the time of the user.” Keywords: search engine, web search engine, semantic search, resource discovery, - advanced search techniques, information retrieval. 1.0 Introduction Retrieval systems in libraries have been historically very efficient and effective as they are strongly supported by cataloging for description and classification systems for organization of information. The same has continued even in the digital era where online catalogs are maintained by library standards such as catalog codes, classification schemes, standard subject headings lists, subject thesauri, etc. However, the information resources in a given library are limited. With the rapid advancement of technology, a large amount of information is being made available on the web in various forms such as text, multimedia, and another format continuously, however, retrieving relevant results from the web search engine is quite difficult.
    [Show full text]
  • NLP-5X Product Brief
    NLP-5x Natural Language Processor With Motor, Sensor and Display Control The NLP-5x is Sensory’s new Natural Language Processor targeting consumer electronics. The NLP-5x is designed to offer advanced speech recognition, text-to-speech (TTS), high quality music and speech synthesis and robotic control to cost-sensitive high volume products. Based on a 16-bit DSP, the NLP-5x integrates digital and analog processing blocks and a wide variety of communication interfaces into a single chip solution minimizing the need for external components. The NLP-5x operates in tandem with FluentChip™ firmware - an ultra- compact suite of technologies that enables products with up to 750 seconds of compressed speech, natural language interface grammars, TTS synthesis, Truly Hands-Free™ triggers, multiple speaker dependent and independent vocabularies, high quality stereo music, speaker verification (voice password), robotics firmware and all application code built into the NLP-5x as a single chip solution. The NLP-5x also represents unprecedented flexibility in application hardware designs. Thanks to the highly integrated architecture, the most cost-effective voice user interface (VUI) designs can be built with as few additional parts as a clock crystal, speaker, microphone, and few resistors and capacitors. The same integration provides all the necessary control functions on-chip to enable cost-effective man-machine interfaces (MMI) with sensing technologies, and complex robotic products with motors, displays and interactive intelligence. Features BROAD
    [Show full text]
  • Google Search by Voice: a Case Study
    Google Search by Voice: A case study Johan Schalkwyk, Doug Beeferman, Fran¸coiseBeaufays, Bill Byrne, Ciprian Chelba, Mike Cohen, Maryam Garret, Brian Strope Google, Inc. 1600 Amphiteatre Pkwy Mountain View, CA 94043, USA 1 Introduction Using our voice to access information has been part of science fiction ever since the days of Captain Kirk talking to the Star Trek computer. Today, with powerful smartphones and cloud based computing, science fiction is becoming reality. In this chapter we give an overview of Google Search by Voice and our efforts to make speech input on mobile devices truly ubiqui- tous. The explosion in recent years of mobile devices, especially web-enabled smartphones, has resulted in new user expectations and needs. Some of these new expectations are about the nature of the services - e.g., new types of up-to-the-minute information ("where's the closest parking spot?") or communications (e.g., "update my facebook status to 'seeking chocolate'"). There is also the growing expectation of ubiquitous availability. Users in- creasingly expect to have constant access to the information and services of the web. Given the nature of delivery devices (e.g., fit in your pocket or in your ear) and the increased range of usage scenarios (while driving, biking, walking down the street), speech technology has taken on new importance in accommodating user needs for ubiquitous mobile access - any time, any place, any usage scenario, as part of any type of activity. A goal at Google is to make spoken access ubiquitously available. We would like to let the user choose - they should be able to take it for granted that spoken interaction is always an option.
    [Show full text]
  • A Guide to Chatbot Terminology
    Machine Language A GUIDE TO CHATBOT TERMINOLOGY Decision Trees The most basic chatbots are based on tree-structured flowcharts. Their responses follow IF/THEN scripts that are linked to keywords and buttons. Natural Language Processing (NLP) A computer’s ability to detect human speech, recognize patterns in conversation, and turn text into speech. Natural Language Understanding A computer’s ability to determine intent, especially when what is said doesn’t quite match what is meant. This task is much more dicult for computers than Natural Language Processing. Layered Communication Human communication is complex. Consider the intricacies of: • Misused phrases • Intonation • Double meanings • Passive aggression • Poor pronunciation • Regional dialects • Subtle humor • Speech impairments • Non-native speakers • Slang • Syntax Messenger Chatbots Messenger chatbots reside within the messaging applications of larger digital platforms (e.g., Facebook, WhatsApp, Twitter, etc.) and allow businesses to interact with customers on the channels where they spend the most time. Chatbot Design Programs There’s no reason to design a messenger bot from scratch. Chatbot design programs help designers make bots that: • Can be used on multiple channels • (social, web, apps) • Have custom design elements • (response time, contact buttons, • images, audio, etc.) • Collect payments • Track analytics (open rates, user • retention, subscribe/unsubscribe • rates) • Allow for human takeover when • the bot’s capabilities are • surpassed • Integrate with popular digital • platforms (Shopify, Zapier, Google • Site Search, etc.) • Provide customer support when • issues arise Voice User Interface A voice user interface (VUI) allows people to interact with a computer through spoken commands and questions. Conversational User Interface Like a voice user interface, a conversational user interface (CUI) allows people to control a computer with speech, but CUI’s dier in that they emulate the nuances of human conversation.
    [Show full text]
  • Voice User Interface on the Web Human Computer Interaction Fulvio Corno, Luigi De Russis Academic Year 2019/2020 How to Create a VUI on the Web?
    Voice User Interface On The Web Human Computer Interaction Fulvio Corno, Luigi De Russis Academic Year 2019/2020 How to create a VUI on the Web? § Three (main) steps, typically: o Speech Recognition o Text manipulation (e.g., Natural Language Processing) o Speech Synthesis § We are going to start from a simple application to reach a quite complex scenario o by using HTML5, JS, and PHP § Reminder: we are interested in creating an interactive prototype, at the end 2 Human Computer Interaction Weather Web App A VUI for "chatting" about the weather Base implementation at https://github.com/polito-hci-2019/vui-example 3 Human Computer Interaction Speech Recognition and Synthesis § Web Speech API o currently a draft, experimental, unofficial HTML5 API (!) o https://wicg.github.io/speech-api/ § Covers both speech recognition and synthesis o different degrees of support by browsers 4 Human Computer Interaction Web Speech API: Speech Recognition § Accessed via the SpeechRecognition interface o provides the ability to recogniZe voice from an audio input o normally via the device's default speech recognition service § Generally, the interface's constructor is used to create a new SpeechRecognition object § The SpeechGrammar interface can be used to represent a particular set of grammar that your app should recogniZe o Grammar is defined using JSpeech Grammar Format (JSGF) 5 Human Computer Interaction Speech Recognition: A Minimal Example const recognition = new window.SpeechRecognition(); recognition.onresult = (event) => { const speechToText = event.results[0][0].transcript;
    [Show full text]
  • VERSE: Bridging Screen Readers and Voice Assistants for Enhanced
    VERSE: Bridging Screen Readers and Voice Assistants for Enhanced Eyes-Free Web Search Alexandra Vtyurina∗ Adam Fourney Meredith Ringel Morris University of Waterloo Microsoft Research Microsoft Research Waterloo, Ontario, Canada Redmond, WA, USA Redmond, WA, USA [email protected] [email protected] [email protected] Leah Findlater Ryen W. White University of Washington Microsoft Research Seattle, WA, USA Redmond, WA, USA [email protected] [email protected] ABSTRACT INTRODUCTION People with visual impairments often rely on screen readers People with visual impairments are often early adopters of when interacting with computer systems. Increasingly, these audio-based interfaces, with screen readers being a prime individuals also make extensive use of voice-based virtual example. Screen readers work by transforming the visual assistants (VAs). We conducted a survey of 53 people who are content in a graphical user interface into audio by vocalizing legally blind to identify the strengths and weaknesses of both on-screen text. To this end, they are an important accessibility technologies, and the unmet opportunities at their intersection. tool for blind computer users – so much so that every major We learned that virtual assistants are convenient and accessible, operating system includes screen reader functionality (e.g., but lack the ability to deeply engage with content (e.g., read VoiceOver1, TalkBack2, Narrator3), and there is a strong mar- beyond the frst few sentences of an article), and the ability ket for third-party offerings (e.g., JAWS4, NVDA5). Despite to get a quick overview of the landscape (e.g., list alternative their importance, screen readers have many limitations.
    [Show full text]
  • Eindversie-Paper-Rianne-Nieland-2057069
    Talking to Linked Data: Comparing voice interfaces for general­purpose data Master thesis of Information Sciences Rianne Nieland Vrije Universiteit Amsterdam [email protected] ABSTRACT ternet access (Miniwatts Marketing Group, 2012) and People in developing countries cannot access informa- 31.1% is literate (UNESCO, 2010). tion on the Web, because they have no Internet access and are often low literate. A solution could be to pro- A solution to the literacy and Internet access problem vide voice-based access to data on the Web by using the is to provide voice-based access to the Internet by us- GSM network. Related work states that Linked Data ing the GSM network (De Boer et al., 2013). 2G mobile could be a useful input source for such voice interfaces. phones are digital mobile phones that use the GSM net- work (Fendelman, 2014). In Africa the primary mode The aim of this paper is to find an efficient way to make of telecommunication is mobile telephony (UNCTAD, general-purpose data, like Wikipedia information, avail- 2007). able using voice interfaces for GSM. To achieve this, we developed two voice interfaces, one for Wikipedia This paper is about how information on the Web ef- and one for DBpedia, by doing requirements elicitation ficiently can be made available using voice interfaces from literature and developing a voice user interface and for GSM. We developed two voice interfaces, one us- conversion algorithms for Wikipedia and DBpedia con- ing Wikipedia and the other using DBpedia. In this cepts. With user tests the users evaluated the two voice paper we see Wikipedia and DBpedia as two different interfaces, to be able to compare them.
    [Show full text]
  • Voice Assistants and Smart Speakers in Everyday Life and in Education
    Informatics in Education, 2020, Vol. 19, No. 3, 473–490 473 © 2020 Vilnius University, ETH Zürich DOI: 10.15388/infedu.2020.21 Voice Assistants and Smart Speakers in Everyday Life and in Education George TERZOPOULOS, Maya SATRATZEMI Department of Applied Informatics, University of Macedonia, Thessaloniki, Greece Email: [email protected], [email protected] Received: November 2019 Abstract. In recent years, Artificial Intelligence (AI) has shown significant progress and its -po tential is growing. An application area of AI is Natural Language Processing (NLP). Voice as- sistants incorporate AI by using cloud computing and can communicate with the users in natural language. Voice assistants are easy to use and thus there are millions of devices that incorporates them in households nowadays. Most common devices with voice assistants are smart speakers and they have just started to be used in schools and universities. The purpose of this paper is to study how voice assistants and smart speakers are used in everyday life and whether there is potential in order for them to be used for educational purposes. Keywords: artificial intelligence, smart speakers, voice assistants, education. 1. Introduction Emerging technologies like virtual reality, augmented reality and voice interaction are reshaping the way people engage with the world and transforming digital experiences. Voice control is the next evolution of human-machine interaction, thanks to advances in cloud computing, Artificial Intelligence (AI) and the Internet of Things (IoT). In the last years, the heavy use of smartphones led to the appearance of voice assistants such as Apple’s Siri, Google’s Assistant, Microsoft’s Cortana and Amazon’s Alexa.
    [Show full text]
  • PDF Preprint
    A Deeper Investigation of the Importance of Wikipedia Links to Search Engine Results NICHOLAS VINCENT, Northwestern University, USA BRENT HECHT, Northwestern University, USA A growing body of work has highlighted the important role that Wikipedia’s volunteer-created content plays in helping search engines achieve their core goal of addressing the information needs of hundreds of millions of people. In this paper, we report the results of an investigation into the incidence of Wikipedia links in search engine results pages (SERPs). Our results extend prior work by considering three U.S. search engines, simulating both mobile and desktop devices, and using a spatial analysis approach designed to study modern SERPs that are no longer just “ten blue links”. We find that Wikipedia links are extremely common in important search contexts, appearing in 67-84% of desktop SERPs for common and trending queries, but less often for medical queries. Furthermore, we observe that Wikipedia links often appear in “Knowledge Panel” SERP elements and are in positions visible to users without scrolling, although Wikipedia appears less often and in less prominent positions on mobile devices. Our findings reinforce the complementary notions that (1) Wikipedia content and research has major impact outside of the Wikipedia domain and (2) powerful technologies like search engines are highly reliant on free content 4 created by volunteers. CCS Concepts: • Human-centered computing → Empirical studies in collaborative and social computing KEYWORDS Wikipedia; search engines; user-generated content; data leverage ACM Reference Format: Nicholas Vincent and Brent Hecht. 2021. A Deeper Investigation of the Importance of Wikipedia Links to Search Engine Results.
    [Show full text]