| Hao Wakat I Akit El Dia De Linha Mitte

Total Page:16

File Type:pdf, Size:1020Kb

| Hao Wakat I Akit El Dia De Linha Mitte |HAO WAKAT I AKITUS009785630B2 EL DIA DE LINHA MITTE (12 ) United States Patent ( 10 ) Patent No. : US 9 , 785 ,630 B2 Willmore et al . (45 ) Date of Patent: Oct . 10, 2017 ( 54 ) TEXT PREDICTION USING COMBINED ( 56 ) References Cited WORD N -GRAM AND UNIGRAM LANGUAGE MODELS U . S . PATENT DOCUMENTS 1 ,559 ,320 A 10 / 1925 Hirsh ( 71 ) Applicant: Apple Inc ., Cupertino , CA (US ) 2 , 180 ,522 A 11/ 1939 Henne (72 ) Inventors : Christopher P . Willmore , Santa Clara , (Continued ) CA (US ) ; Nicholas K . Jong, Seattle , FOREIGN PATENT DOCUMENTS WA (US ) ; Justin S . Hogg , San Francisco , CA (US ) CA 2666438 C 6 /2013 CH 681573 A5 4 / 1993 (73 ) Assignee : Apple Inc ., Cupertino , CA (US ) ( Continued ) ( * ) Notice : Subject to any disclaimer , the term of this OTHER PUBLICATIONS patent is extended or adjusted under 35 Carpendale et al. , “ 3 -Dimensional Pliable Surfaces: For the Effec U . S . C . 154 (b ) by 2 days. tive Presentation of Visual Information ” , UIST ' 95 Proceedings of the 8th Annual ACM Symposium on User Interface and Software (21 ) Appl . No. : 14/ 724, 641 Technology, Nov. 14 - 17 , 1995 , pp . 217 - 226 . May 28 , 2015 (Continued ) (22 ) Filed : Primary Examiner — Anne Thomas -Homescu (74 ) Attorney , Agent, or Firm — Morrison & Foerster (65 ) Prior Publication Data LLP US 2015 /0347383 A1 Dec . 3 , 2015 (57 ) ABSTRACT Systems and processes are disclosed for predicting words in a text entry environment. Candidate words and probabilities Related U . S . Application Data associated therewith can be determined by combining a (60 ) Provisional application No .62 /005 , 942 , filed on May word n - gram language model and a unigram language 30 , 2014 . model. Using the word n - gram language model , based on previously entered words, candidate words can be identified and a probability can be calculated for each candidate word . (51 ) Int . Ci. Using the unigram language model, based on a character G06F 17 / 20 ( 2006 . 01 ) entered for a new word , candidate words beginning with the G06F 17 /28 ( 2006 .01 ) character can be identified along with a probability for each ( Continued ) candidate word . In some examples , a geometry score can be (52 ) U . S. Ci. included in the unigram probability related to typing geom CPC .. .. G06F 17 /276 ( 2013 . 01 ) etry on a virtual keyboard . The probabilities of the n - gram (58 ) Field of Classification Search language model and unigram model can be combined , and CPC .. GO6F 3 /04886 ; GO6F 3 /0485 ; GO6F 3 /0237 ; the candidate word or words having the highest probability G06F 3 /01 : G06F 9 /4443 : can be displayed for a user. ( Continued ) 20 Claims , 5 Drawing Sheets Process 200 Receive typed input from a user Using a word n - gram model, determine a first probability of a predicted word based on anL 2204 14 a previously entered word Using a unigram model, determine a second probability of the predicted word based on a inom 206 prefix character of a new word Determine a combined probability of the predicted word based on the first probability Consom 208 and the second probability Cause the predicted word to be displayed . 210 based on the combined probability US 9 ,785 , 630 B2 Page 2 (51 ) Int. CI. 4 ,677 , 570 A 6 / 1987 Taki GIOL 15 /00 ( 2013 .01 ) 4 ,680 ,429 A 7 / 1987 Murdock et al. 4 ,680 , 805 A 7 / 1987 Scott GIOL 15 / 04 ( 2013 .01 ) 4 ,686 , 522 A 8 /1987 Hernandez et al . GIOL 15 /06 ( 2013 .01 ) 4 ,688 , 195 A 8 / 1987 Thompson et al. GIOL 15 / 14 ( 2006 . 01 ) 4 ,692 , 941 A 9 / 1987 Jacks et al. 4 ,698 ,625 A 10 / 1987 McCaskill et al. GIOL 15 / 18 ( 2013 . 01 ) 4 ,709 , 390 A 11 / 1987 Atal et al. GIOL 17 /00 ( 2013 .01 ) 4 ,713 ,775 A 12 / 1987 Scott et al . G06F 3 /00 ( 2006 .01 ) 4 ,718 ,094 A 1 / 1988 Bahl et al . G06F 3 /048 ( 2013 .01 ) 4 , 724 , 542 A 2 / 1988 Williford 4 , 726 ,065 A 2 / 1988 Froessl G06F 17 / 00 ( 2006 .01 ) 4 ,727 , 354 A 2 / 1988 Lindsay G06F 17 / 27 ( 2006 . 01 ) RE32 ,632 E 3 / 1988 Atkinson GOOF 17 / 21 ( 2006 .01 ) 4 ,736 , 296 A 4 / 1988 Katayama et al. G06F 17 / 22 ( 2006 .01 ) 4 ,750 , 122 A 6 / 1988 Kaji et al. G06F 17 / 24 ( 2006 . 01 ) 4 ,754 ,489 A 6 / 1988 Bokser 4 , 755 , 811 A 7 / 1988 Slavin et al . G06F 17 / 25 ( 2006 .01 ) 4 ,776 ,016 A 10 / 1988 Hansen G06F 17 / 26 ( 2006 .01 ) 4 ,783 , 804 A 11/ 1988 Juang et al . GOOF 17/ 30 ( 2006 .01 ) 4 ,783 , 807 A 11/ 1988 Marley G06F 3 / 02 ( 2006 .01 ) 4 , 785 ,413 A 11/ 1988 Atsumi 4 ,790 ,028 A 12 / 1988 Ramage G09G 5 / 00 ( 2006 .01 ) 4 , 797 , 930 A 1 / 1989 Goudie (58 ) Field of Classification Search 4 , 802 , 223 A 1 / 1989 Lin et al . CPC .. G06F 17 /248 ; GO6F 1 /626 ; G06F 17 /2785 ; 4 , 803 ,729 A 2 / 1989 Baker G06F 17 / 2735 ; G06F 17 / 274 ; G06F 4 ,807 , 752 A 2 / 1989 Chodorow 17 / 289 ; GOOF 17 / 02 ; G06F 17 /005 ; G06F 4 ,811 , 243 A 3 / 1989 Racine 4 , 813 , 074 A 3 / 1989 Marcus 9 /4448 ; G10L 17 /02 ; GIOL 17 /005 ; G10L 4 ,819 ,271 A 4 / 1989 Bahl et al. 15 /08 ; G10L 15 / 063 ; G10L 15 / 265 ; GIOL 4 , 827 , 518 A 5 / 1989 Feustel et al. 15 /22 ; G10L 15 / 187 ; G10L 15 / 142 4 , 827 , 520 A 5 / 1989 Zeinstra See application file for complete search history . 4 , 829 , 576 A 5 / 1989 Porter 4 , 829 , 583 A 5 / 1989 Monroe et al . 4 , 831 ,551 A 5 / 1989 Schalk et al . (56 ) References Cited 4 ,833 ,712 A 5 / 1989 Bahl et al . 4 , 833 , 718 A 5 / 1989 Sprague U . S . PATENT DOCUMENTS 4 , 837 ,798 A 6 / 1989 Cohen et al. 4 ,837 , 831 A 6 / 1989 Gillick et al. 3 , 704 , 345 A 11 / 1972 Coker et al. 4 , 839 , 853 A 6 / 1989 Deerwester et al. 3 ,710 ,321 A 1 / 1973 Rubenstein 4 , 852 , 168 A 7 / 1989 Sprague 3 , 828 , 132 A 8 / 1974 Flanagan et al. 4 ,862 , 504 A 8 / 1989 Nomura 3 , 979 ,557 A 9 / 1976 Schulman et al . 4 ,875 , 187 A 10 / 1989 Smith 4 ,013 ,085 A 3 / 1977 Wright 4 , 878 , 230 A 10 / 1989 Murakami et al . 4 ,081 ,631 A 3 / 1978 Feder 4 , 887 , 212 A 12 / 1989 Zamora et al . 4 ,090 ,216 A 5 / 1978 Constable 4 ,896 , 359 A 1 / 1990 Yamamoto et al. 4 , 107 , 784 A 8 / 1978 Van Bemmelen 4 , 903 ,305 A 2 / 1990 Gillick et al. 4 , 108 , 211 A 8 / 1978 Tanaka 4 , 905 , 163 A 2 / 1990 Garber et al. 4 , 159 , 536 A 6 / 1979 Kehoe et al . 4 , 908 , 867 A 3 / 1990 Silverman 4 , 181 , 821 A 1 / 1980 Pirz et al . 4 , 914 , 586 A 4 / 1990 Swinehart et al . 4 , 204 ,089 A 5 / 1980 Key et al . 4 , 914 , 590 A 4 / 1990 Loatman et al. 4 , 241 , 286 A 12 / 1980 Gordonvis 4 ,918 , 723 A 4 / 1990 Iggulden et al . 4 ,253 ,477 A 3 / 1981 Eichman 4 ,926 ,491 A 5 / 1990 Maeda et al. 4 ,278 , 838 A 7 / 1981 Antonov 4 , 928 , 307 A 5 / 1990 Lynn 4 ,282 , 405 A 8 / 1981 Taguchi 4 ,931 , 783 A 6 / 1990 Atkinson 4 , 310 ,721 A 1 / 1982 Manley et al. 4 ,935 ,954 A 6 / 1990 Thompson et al. 4 , 332 , 464 A 6 / 1982 Bartulis et al . 4 , 939 ,639 A 7 / 1990 Lee et al . 4 , 348 , 553 A 9 / 1982 Baker et al. 4 , 941 , 488 A 7 / 1990 Marxer et al. 4 , 384 , 169 A 5 / 1983 Mozer et al. 4 ,944 ,013 A 7 / 1990 Gouvianakis et al. 4 , 386 , 345 A 5 / 1983 Narveson et al . 4 ,945 , 504 A 7 / 1990 Nakama et al. 4 ,433 , 377 A 2 / 1984 Eustis et al . 4 ,953 , 106 A 8 / 1990 Gansner et al . 4 ,451 , 849 A 5 / 1984 Fuhrer 4 , 955 ,047 A 9 / 1990 Morganstein et al . 4 ,485 , 439 A 11/ 1984 Rothstein 4 , 965 , 763 A 10 / 1990 Zamora 4 ,495 ,644 A 1 / 1985 Parks et al. 4 ,972 ,462 A 11/ 1990 Shibata 4 ,513 ,379 A 4 / 1985 Wilson et al . 4 ,974 , 191 A 11/ 1990 Amirghodsi et al . 4 , 513 , 435 A 4 / 1985 Sakoe et al. 4 , 975 , 975 A 12 / 1990 Filipski 4 , 555 , 775 A 11/ 1985 Pike 4 , 977 , 598 A 12 / 1990 Doddington et al . 4 , 577 , 343 A 3 / 1986 Oura 4 ,980 , 916 A 12 / 1990 Zinser 4 ,586 , 158 A 4 / 1986 Brandle 4 , 985 , 924 A 1 / 1991 Matsuura 4 , 587 ,670 A 5 / 1986 Levinson et al. 4 , 589 , 022 A 5 / 1986 Prince et al . 4 , 992 , 972 A 2 / 1991 Brooks et al. 4 ,611 , 346 A 9 / 1986 Bednar et al . 4 , 994 , 966 A 2 / 1991 Hutchins 4 ,615 , 081 A 10 / 1986 Lindahl 4 , 994 , 983 A 2 / 1991 Landell et al . 4 ,618 , 984 A 10 / 1986 Das et al . 5 , 001 ,774 A 3 / 1991 Lee 4 ,642 ,790 A 2 / 1987 Minshull et al. 5 ,003 , 577 A 3 / 1991 Ertz et al . 4 ,653 ,021 A 3 / 1987 Takagi 5 ,007 ,095 A 4 / 1991 Nara et al . 4 ,654 , 875 A 3 / 1987 Srihari et al .
Recommended publications
  • A Multilingual Keyboard and Mouse Interface for Motor-Impaired Users
    SUITEDasher – A Multilingual Keyboard and Mouse Interface for Motor-Impaired Users David Lyle and Bill Manaris Computer Science Department, College of Charleston 66 George Street, Charleston, SC 29424, USA [email protected], [email protected] Abstract This paper presents the design of SUITEDasher – an open-source, multilingual speech user interface for motor- impaired users. SUITEDasher’s architecture is based on SUITEKeys – a speech user interface for manipulating a speech keyboard and mouse. SUITEDasher has three main design objectives: (a) to improve on the usability of its predecessor; (b) to provide for multilingual access; and (c) to be platform independent. To achieve these objectives, SUITEDasher employs a minimal graphical user interface, and incorporates a trigram-based probabilistic model from Dasher – a text-entry interface driven by pointing gestures. Multilingual access is achieved through a set of syntactic, lexical, and (potentially) phonetic models that may be loaded dynamically at run time. Finally, SUITEDasher is being implemented in Java as a cross-platform (Windows, Mac OS X, and Linux) application. While refining the design, we have developed several prototypes, the latest of which has been implemented in Visual Basic and C++ on the Windows platform. Preliminary usability results indicate that, in terms of typing speed, SUITEDasher appears to be 30% faster than its predecessor, and as fast as Dasher. Keywords: Universal access, motor impairments, speech user interfaces, listening keyboard, usability evaluation. 1 Introduction Several studies have been performed exploring the effectiveness of speech as an alternative to the physical keyboard for alphanumeric data entry. Murray et al. (1983) found that, in the context of controlling cursor movement, keyboard input is twice as fast and more preferred by users compared to speech.
    [Show full text]
  • 7 35138412 1.Pdf (2.296Mb)
    MAUU(D)5900 MASTER THESIS in Universal Design of ICT October 2018 An Accessible Directions-based Text Entry Method Using Two-thumb Touch Typing Linghui Ye Department of Computer Science Faculty of Technology, Art and Design Master Thesis Phase III Report Contents Abstract ......................................................................................................................................................... 3 1. Introduction .............................................................................................................................................. 5 2. Related work ............................................................................................................................................. 8 3. The prototype ......................................................................................................................................... 15 3.1 Physical direction .............................................................................................................................. 20 3.2 Resolving ambiguities ....................................................................................................................... 23 3.3 Special characters ............................................................................................................................. 23 4. Methodology ........................................................................................................................................... 25 4.1 Experimental design .........................................................................................................................
    [Show full text]
  • Fast and Precise Touch-Based Text Entry for Head-Mounted Augmented Reality with Variable Occlusion
    1 Fast and Precise Touch-Based Text Entry for Head-Mounted Augmented Reality with Variable Occlusion JOHN J. DUDLEY, University of Cambridge, United Kingdom KEITH VERTANEN, Michigan Technological University, USA PER OLA KRISTENSSON, University of Cambridge, United Kingdom We present the VISAR keyboard: an augmented reality (AR) head-mounted display (HMD) system that supports text entry via a virtualised input surface. Users select keys on the virtual keyboard by imitating the process of single-hand typing on a physical touchscreen display. Our system uses a statistical decoder to infer users’ intended text and to provide error-tolerant predictions. There is also a high-precision fall-back mechanism to support users in indicating which keys should be unmodified by the auto-correction process. A unique advantage of leveraging the well-established touch input paradigm is that our system enables text entry with minimal visual clutter on the see-through display, thus preserving the user’s field-of-view. We iteratively designed and evaluated our system and show that the final iteration of the system supports a mean entry rate of 17.75 wpm with a mean character error rate less than 1%. This performance represents a 19.6% improvement relative to the state-of-the-art baseline investigated: a gaze-then-gesture text entry technique derived from the system keyboard on the Microsoft HoloLens. Finally, we validate that the system is effective in supporting text entry in a fully mobile usage scenario likely to be encountered in industrial applications of AR HMDs. CCS Concepts: • Human-centered computing → Text input; Additional Key Words and Phrases: augmented reality, text entry 1 INTRODUCTION Recent progress in head-mounted displays (HMDs) for augmented reality (AR), such as the Microsoft HoloLens, demonstrates the commercial potential of AR to support new forms of interaction and work in a range of industries including construction, education and health.
    [Show full text]
  • Digital Access Project
    DIGITAL ACCESS PROJECT LEARNING MODULE Computer hardware, basic digital terminology and typing skills Code: M5BO Prepared by: ITPIO October 2018 This project has been funded with support from the European Commission. This publication reflects the views only of the authors, and the Commission cannot be held responsible for any use which may be made of the information contained therein. Contents SUMMARY ..............................................................................................................................5 KEYWORDS .............................................................................................................................5 MODULE OBJECTIVES .............................................................................................................5 UNIT 1: USING THE INTERFACE OF COMPUTERS .....................................................................7 Specific objectives of Unit 1 ................................................................................................7 Display ................................................................................................................................7 Mouse and Touchpad .........................................................................................................8 Keyboard ............................................................................................................................9 Printer...............................................................................................................................11
    [Show full text]
  • QB-Gest: Qwerty Bimanual Gestural Input for Eyes-Free Smartphone Text Input
    QB-Gest: Qwerty Bimanual Gestural Input for Eyes-Free Smartphone Text Input Linghui Ye1, Frode Eika Sandnes1,2(&) , and I. Scott MacKenzie3 1 Department of Computer Science, Oslo Metropolitan University, 0130 Oslo, Norway [email protected], [email protected] 2 Institute of Technology, Kristiania University College, 0153 Oslo, Norway 3 Department of Computer Science, York University, Toronto, ON M3J 1P3, Canada [email protected] Abstract. We developed QB-Gest, a bimanual text entry method based on simple gestures where users drag their thumbs in the direction of the desired letter while visualizing the Qwerty-layout. In an experiment with four sessions of testing, 20 users achieved text entry rates of 11.1 wpm eyes-free and 14.1 wpm eyes-on. An expert user achieved an eyes-free rate of 24.9 wpm after 10 rounds of entering the-quick-brown-fox phrase. The method holds potential for users with low vision and certain types of reduced motor function. Keywords: Mobile text entry Á Eyes free text Á Gestures Á Qwerty 1 Introduction Smartphones have become an important tool in modern society by facilitating com- munication independent of time and place. Many smartphone tasks require text input, such as searching the web, sending emails, or messaging. Text input typically uses a default virtual keyboard. Yet, many users find it hard to use virtual smartphone key- boards compared to physical desktop keyboards [1], because input requires accurately hitting keys without tactile feedback. Virtual smartphone keys are smaller than physical desktop keys; so, input is both visually intensive and requires careful eye-motor coordination [2].
    [Show full text]
  • 1 Background Many Applications for Handheld Electronic Devices Depend on the User Being Able to Enter Text
    ABSTRACT When using your smartphone (or similar handheld electronic device) with one hand, you have to type with the thumb, which isn’t the finger you would normally use. The more agile fingers are tied up with holding the device. This is due to the interface concept, not to any functional deficit of the human hand. You could operate these devices with the use of all fingers, and not just for typing, but that calls for a different interface paradigm, one that mirrors the ergonomic capabilities of the human hand. You engage this kind of interface where your fingers spontaneously come into contact with the device when you grasp it, not at predetermined locations that are not necessarily easy to reach. This interface dynamically maps the controls of the device to the contact patches of your fingers and generates a visual, audible or haptic tag to tell you what function a finger controls. To activate a control you engage it with the corresponding finger. KEYWORDS Topical† tactile interfaces, atopical* tactile interfaces, operative interaction with back and sides of device, accessibility issues with virtual controls, atypical handheld devices (disks, cylinders, torons etc.) †characterised by operative interactions confined to defined locations *characterised by operative interactions not confined to defined locations 1 Background Many applications for handheld electronic devices depend on the user being able to enter text. To this end, the tactile user interfaces of these devices are configured with a miniature keyboard. The idea behind this is presumably to replicate the functionality of a standard keyboard. This may be aiming a little too high, though.
    [Show full text]
  • Interactions with Smartphones and Smartwatches: Context-Awareness, Text Entry Interfaces, and Input Beyond Touch
    INTERACTIONS WITH SMARTPHONES AND SMARTWATCHES: CONTEXT-AWARENESS, TEXT ENTRY INTERFACES, AND INPUT BEYOND TOUCH Rajkumar Darbar INTERACTIONS WITH SMARTPHONES AND SMARTWATCHES: CONTEXT-AWARENESS, TEXT ENTRY INTERFACES, AND INPUT BEYOND TOUCH Thesis submitted to the Indian Institute of Technology Kharagpur for award of the degree of Master of Science (MS) - by Research by Rajkumar Darbar Under the guidance of Dr. Debasis Samanta Computer Science and Engineering Indian Institute of Technology Kharagpur Kharagpur - 721 302, India July 2016 ⃝c 2016 Rajkumar Darbar. All rights reserved. CERTIFICATE OF APPROVAL 00/00/0000 Certified that the thesis entitled Interactions with Smartphones and Smartwatches: Context-Awareness, Text Entry Interfaces, and Input Beyond Touch submit- ted by Rajkumar Darbar to the Indian Institute of Technology, Kharagpur, for the award of the degree Master of Science has been accepted by the external examiners and that the student has successfully defended the thesis in the viva-voce examination held today. (Member of DAC) (Member of DAC) (Member of DAC) (Supervisor) (Internal Examiner) (Chairman) CERTIFICATE This is to certify that the thesis entitled Interactions with Smartphones and Smart- watches: Context-Awareness, Text Entry Interfaces, and Input Beyond Touch submitted by Rajkumar Darbar to Indian Institute of Technology Kharagpur, is a record of bona fide research work under my supervision and I consider it worthy of consideration for the award of the degree of Master of Science (by Research) of the Institute. Date: 20/07/2016 Dr. Debasis Samanta Associate Professor Computer Science and Engineering Indian Institute of Technology Kharagpur Kharagpur - 721 302, India DECLARATION I certify that a. The work contained in the thesis is original and has been done by myself under the general supervision of my supervisor.
    [Show full text]
  • 30 Fast and Precise Touch-Based Text Entry for Head-Mounted
    30 Fast and Precise Touch-Based Text Entry for Head-Mounted Augmented Reality with Variable Occlusion JOHN J. DUDLEY, University of Cambridge KEITH VERTANEN, Michigan Technological University PER OLA KRISTENSSON, University of Cambridge We present the VISAR keyboard: An augmented reality (AR) head-mounted display (HMD) system that sup- ports text entry via a virtualised input surface. Users select keys on the virtual keyboard by imitating the process of single-hand typing on a physical touchscreen display. Our system uses a statistical decoder to infer users’ intended text and to provide error-tolerant predictions. There is also a high-precision fall-back mechanism to support users in indicating which keys should be unmodified by the auto-correction process. A unique advantage of leveraging the well-established touch input paradigm is that our system enables text entry with minimal visual clutter on the see-through display, thus preserving the user’s field-of-view. We it- eratively designed and evaluated our system and show that the final iteration of the system supports a mean entry rate of 17.75wpm with a mean character error rate less than 1%. This performance represents a 19.6% improvement relative to the state-of-the-art baseline investigated: A gaze-then-gesture text entry technique derived from the system keyboard on the Microsoft HoloLens. Finally, we validate that the system is effective in supporting text entry in a fully mobile usage scenario likely to be encountered in industrial applications of AR HMDs. CCS Concepts: • Human-centered computing → Text input; Additional Key Words and Phrases: Augmented reality, text entry ACM Reference format: John J.
    [Show full text]
  • ( 12 ) United States Patent ( 10 ) Patent No.: US 10,474,347 B2 Tortosa Et Al
    US010474347B2 ( 12 ) United States Patent ( 10 ) Patent No.: US 10,474,347 B2 Tortosa et al. (45 ) Date of Patent : Nov. 12 , 2019 (54 ) AUTOMATED MODIFICATION OF 2012/0036468 A1 * 2/2012 Colley G06F 3/0418 GRAPHICAL USER INTERFACES 715/773 2012/0169613 A1 * 7/2012 Armstrong GO6F 3/04886 345/173 ( 71) Applicant: International Business Machines 2013/0019191 A1 * 1/2013 Arnold G06F 3/04886 Corporation , Armonk , NY (US ) 715/765 2013/0191779 Al 7/2013 Radakovitz et al . ( 72 ) Inventors : Andrea Tortosa , Rome (IT ) ; Francesco Maria Carteri, Rome (IT ) ; Massimo (Continued ) Marra , Rome (IT ) FOREIGN PATENT DOCUMENTS ( 73 ) Assignee : International Business Machines WO 2012127101 A1 9/2012 Corporation , Armonk , NY (US ) ( * ) Notice: Subject to any disclaimer , the term of this OTHER PUBLICATIONS patent is extended or adjusted under 35 Reynaga et al ., “ Exploring the Usability of CAPTCHAS on U.S.C. 154 ( b ) by 542 days . Smartphones: Comparisons and Recommendations” , USEC '15 , Feb. 8 , 2015 , San Diego, CA , USA , pp . 1-10 , Copyright 2015 ( 21) Appl. No .: 14 /918,633 Internet Society , ISBN 1-891562-40-1 , < http://dx.doi.org/10.14722/ usec.2015.23006 > . ( 22 ) Filed : Oct. 21 , 2015 (Continued ) (65 ) Prior Publication Data Primary Examiner Andrew T Chiusano US 2017/0115794 A1 Apr. 27 , 2017 (74 ) Attorney, Agent, or Firm — Dmitry Paskalov (51 ) Int . Cl. ( 57 ) ABSTRACT G06F 3/048 ( 2013.01 ) Information that describes one or more touch events per G06F 3/0488 ( 2013.01) formed on a touchscreen display to interact with a graphical G06F 9/451 (2018.01 ) user interface (GUI ) of a software application is received .
    [Show full text]
  • Facilitating Text Entry on Smartphones with QWERTY Keyboard for Users with Parkinson’S Disease
    Facilitating Text Entry on Smartphones with QWERTY Keyboard for Users with Parkinson’s Disease Yuntao Wang Ao Yu Xin Yi∗ Department of Computer Science and Global Innovation Exchange (GIX), Department of Computer Science and Technology, Key Laboratory of Tsinghua University and University of Technology, Key Laboratory of Pervasive Computing, Ministry of Washington, 98005, WA, USA Pervasive Computing, Ministry of Education, Tsinghua University, [email protected] Education, Tsinghua University, Beijing, China, 100084 Beijing, China, 100084 [email protected] [email protected] Yuanwei Zhang Ishan Chatterjee Shwetak Patel Global Innovation Exchange (GIX), Paul G. Allen School of Computer Paul G. Allen School of Computer Tsinghua University and University of Science and Engineering, University Science and Engineering, University Washington, 98005, WA, USA of Washington, 98195, Seattle, USA of Washington, 98195, Seattle, USA [email protected] [email protected] [email protected] Yuanchun Shi Department of Computer Science and Technology, Key Laboratory of Pervasive Computing, Ministry of Education, Tsinghua University, Beijing, China, 100084 [email protected] ABSTRACT CCS CONCEPTS QWERTY is the primary smartphone text input keyboard confgu- • Human-centered computing ! Text input; Accessibility. ration. However, insertion and substitution errors caused by hand tremors, often experienced by users with Parkinson’s disease, can KEYWORDS severely afect typing efciency and user experience. In this paper, Parkinson’s disease, text entry, QWERTY keyboard, touch model, we investigated Parkinson’s users’ typing behavior on smartphones. statistical decoding In particular, we identifed and compared the typing characteristics generated by users with and without Parkinson’s symptoms. We ACM Reference Format: then proposed an elastic probabilistic model for input prediction.
    [Show full text]
  • Text Blaster: a Multi-Player Touchscreen Typing Game
    Text Blaster: A Multi-Player Touchscreen Typing Game Figure 1: A group playing Text Keith Vertanen Per Ola Kristensson Abstract Blaster. Each player uses an Montana Tech University of St Andrews Text Blaster is a multi-player shoot 'em up game based on Android mobile device to control Butte, Montana USA St Andrews, UK players typing sentences on a mobile device's touchscreen their ship. The global state of [email protected] [email protected] keyboard. Players attempt to be the last player standing the game is shown on the large by using the speed, precision, and timing of their typing to projection wall. annihilate competing players. Our game utilizes a Justin Emge sentence-based decoding approach in which users type an Montana Tech entire sentence before our auto-correction algorithm infers Butte, Montana USA the most likely text. Text Blaster provides an engaging [email protected] and competitive game for use in investigating performance and design aspects of touchscreen text entry interfaces. Haythem Memmi Author Keywords Montana Tech Text entry; game design; touchscreen keyboard Butte, Montana USA [email protected] ACM Classification Keywords H.5.2 [Information interfaces and presentation: User Interfaces]: Input devices and strategies. Introduction Permission to make digital or hard copies of part or all of this work for In this paper we describe Text Blaster, a multi-player personal or classroom use is granted without fee provided that copies are not game that we created based around users typing sentences made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page.
    [Show full text]
  • Ye.Pdf (2.296Mb)
    MAUU(D)5900 MASTER THESIS in Universal Design of ICT October 2018 An Accessible Directions-based Text Entry Method Using Two-thumb Touch Typing Linghui Ye Department of Computer Science Faculty of Technology, Art and Design Master Thesis Phase III Report Contents Abstract ......................................................................................................................................................... 3 1. Introduction .............................................................................................................................................. 5 2. Related work ............................................................................................................................................. 8 3. The prototype ......................................................................................................................................... 15 3.1 Physical direction .............................................................................................................................. 20 3.2 Resolving ambiguities ....................................................................................................................... 23 3.3 Special characters ............................................................................................................................. 23 4. Methodology ........................................................................................................................................... 25 4.1 Experimental design .........................................................................................................................
    [Show full text]