Direct Pen Interaction with a Conventional Graphical User Interface
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Direct Interaction with Large Displays Through Monocular Computer Vision
DIRECT INTERACTION WITH LARGE DISPLAYS THROUGH MONOCULAR COMPUTER VISION A thesis submitted in fulfilment of the requirements for the degree of Doctor of Philosophy in the School of Information Technologies at The University of Sydney Kelvin Cheng October 2008 © Copyright by Kelvin Cheng 2008 All Rights Reserved Abstract Large displays are everywhere, and have been shown to provide higher productivity gain and user satisfaction compared to traditional desktop monitors. The computer mouse remains the most common input tool for users to interact with these larger displays. Much effort has been made on making this interaction more natural and more intuitive for the user. The use of computer vision for this purpose has been well researched as it provides freedom and mobility to the user and allows them to interact at a distance. Interaction that relies on monocular computer vision, however, has not been well researched, particularly when used for depth information recovery. This thesis aims to investigate the feasibility of using monocular computer vision to allow bare-hand interaction with large display systems from a distance. By taking into account the location of the user and the interaction area available, a dynamic virtual touchscreen can be estimated between the display and the user. In the process, theories and techniques that make interaction with computer display as easy as pointing to real world objects is explored. Studies were conducted to investigate the way human point at objects naturally with their hand and to examine the inadequacy in existing pointing systems. Models that underpin the pointing strategy used in many of the previous interactive systems were formalized. -
Motion and Context Sensing Techniques for Pen Computing
Motion and Context Sensing Techniques for Pen Computing Ken Hinckley1, Xiang ‘Anthony’ Chen1,2, and Hrvoje Benko1 * Microsoft Research, Redmond, WA, USA1 and Carnegie Mellon University Dept. of Computer Science2 ABSTRACT We explore techniques for a slender and untethered stylus prototype enhanced with a full suite of inertial sensors (three-axis accelerometer, gyroscope, and magnetometer). We present a taxonomy of enhanced stylus input techniques and consider a number of novel possibilities that combine motion sensors with pen stroke and touchscreen inputs on a pen + touch slate. These Fig. 1 Our wireless prototype has accelerometer, gyro, and inertial sensors enable motion-gesture inputs, as well sensing the magnetometer sensors in a ~19 cm Χ 11.5 mm diameter stylus. context of how the user is holding or using the stylus, even when Our system employs a custom pen augmented with inertial the pen is not in contact with the tablet screen. Our initial results sensors (accelerometer, gyro, and magnetometer, each a 3-axis suggest that sensor-enhanced stylus input offers a potentially rich sensor, for nine total sensing dimensions) as well as a low-power modality to augment interaction with slate computers. radio. Our stylus prototype also thus supports fully untethered Keywords: Stylus, motion sensing, sensors, pen+touch, pen input operation in a slender profile with no protrusions (Fig. 1). This allows us to explore numerous interactive possibilities that were Index Terms: H.5.2 Information Interfaces & Presentation: Input cumbersome in previous systems: our prototype supports direct input on tablet displays, allows pen tilting and other motions far 1 INTRODUCTION from the digitizer, and uses a thin, light, and wireless stylus. -
Microsoft Publisher Version 2007
Microsoft Publisher version 2007 Getting Started Select a publication type from the left-hand panel. Your design choices will appear in the center panel. Page 1 of 8 © D. Rich, 2010 Single click a design choice from the center panel. In the right-hand panel, you can edit the default color and font scheme if desired. Options will differ, depending on the chosen publication type. Then click the Create button in the lower right-hand corner. Publication Options NOTE: available options depend on publication type Page 2 of 8 © D. Rich, 2010 To close Task Pane, click X. To reopen at any time, choose View > Task Pane. Zoom Use drop down menu or + - icons. Multi-page publications contain navigation buttons at the bottom of the screen. Page 3 of 8 © D. Rich, 2010 Working With Text Editing Text All text in Publisher is contained within text boxes. These text boxes, by default, contain “filler text.” You will need to delete the filler text and enter your own text. Formatting Text All text in Publisher is contained within text boxes. You will need to click inside the text box to select the text you wish to edit. Edit and format as you would in Word, by using the Formatting toolbar, You can also choose Format > Font for more formatting options. Text Boxes If text is too large for text box, the text box overflow (A…) symbol appears. To enlarge a text box, grab and drag one of the white circles. To move a text box, hover your mouse on the edge of the box and the crosshairs icon will appear. -
Designing and Deploying an Information Awareness Interface
Designing and Deploying an Information Awareness Interface JJ Cadiz, Gina Venolia, Gavin Jancke, Anoop Gupta Collaboration & Multimedia Systems Team Microsoft Research One Microsoft Way, Redmond, WA 98052 {jjcadiz; ginav; gavinj; anoop} @microsoft.com ABSTRACT appeal to a broader audience. Although the ideas and The concept of awareness has received increasing attention lessons generated by such prototypes are valuable, they over the past several CSCW conferences. Although many leave a critical question: Why did these prototypes fail to awareness interfaces have been designed and studied, most provide users with substantial value relative to cost? What have been limited deployments of research prototypes. In combination of features, design, and process will help an this paper we describe Sideshow, a peripheral awareness application succeed in establishing a healthy user interface that was rapidly adopted by thousands of people in population? our company. Sideshow provides regularly updated Sideshow started as one more idea for an interface designed peripheral awareness of a broad range of information from to provide users with peripheral awareness of important virtually any accessible web site or database. We discuss information. Rather than concentrate on a specific Sideshow’s design and the experience of refining and awareness issue, the research team set out to incorporate a redesigning the interface based on feedback from a rapidly range of features into a versatile and extensible system for expanding user community. dynamic information awareness that could be easily Keywords deployed, extended by third parties, and quickly evolved in Situational awareness, peripheral awareness, awareness, response to users’ experiences. computer mediated communication, information overload What happened was something akin to an epidemic within 1 INTRODUCTION our company. -
Using the Tools Palette in Collaborate (The Whiteboard)
Using the Tools Palette in Collaborate (The Whiteboard) The Tools Palette contains drawing and text tools that can be used to create and manipulate objects on the Whiteboard. It also has options to place a screenshot or clip art. The Tools Palette Tools in the palette with a tabbed lower right corner have additional optional tools that can be swapped into their spot. Click and hold the tool to expand the options. Participants have access to the Whiteboard tools if they have been granted the Whiteboard tools permission. The Properties Editor When you click on any of the text or drawing tools in the Tools palette, the Properties Editor palette will appear at the bottom of the Whiteboard. From here you can format your text or graphic by specifying various properties. The contents of the Properties Editor palette changes depending on which tool is being used or which objects are selected. 1 Updated November 29, 2012 Inserting Text The Whiteboard has two text tools: The Simple Text Tool ( ) creates text as an object directly on the canvas. The text can be manipulated just like a shape or image in a graphic design program. The Text Editor ( ) creates a bounding box for text, and the text inside can be edited as in a word processor or publishing program. To insert text on to the Whiteboard: 1. Click on the Text Editor Tool. If the Text Editor tool is not in the Tools Palette, click and hold the Simple Text tool until it expands to reveal the Text Editor tool icon, then click on the Text Editor tool. -
Bforartists UI Redesign Design Document Part 2 - Theming
Bforartists UI redesign Design document part 2 - Theming Content Preface...........................................................................................................................6 The editor and window types......................................................................................7 Python console.............................................................................................................8 Layout:................................................................................................................................................................8 The Console Window.........................................................................................................................................8 Menu bar with a menu........................................................................................................................................8 Dropdown box with icon....................................................................................................................................9 RMB menu for menu bar....................................................................................................................................9 Toolbar................................................................................................................................................................9 Button Textform..................................................................................................................................................9 -
User Interface Aspects of a Human-Hand Simulation System
User Interface Aspects of a Human-Hand Simulation System Beifang YI Department of Computer Science and Engineering University of Nevada, Reno Reno, NV, 89557, USA Frederick C. HARRIS, Jr. Department of Computer Science and Engineering University of Nevada, Reno Reno, NV, 89557, USA Sergiu M. DASCALU Department of Computer Science and Engineering University of Nevada, Reno Reno, NV, 89557, USA Ali EROL Department of Computer Science and Engineering University of Nevada, Reno Reno, NV, 89557, USA ABSTRACT in a virtual environment is the base of a graphical user interface (GUI), which we call Virtual Hand. This paper describes the user interface design for a human-hand simulation system, a virtual environment that produces ground Virtual Hand simulates the human hand motion by modifying truth data (life-like human hand gestures and animations) and the angular parameters (i.e., finger joint angles) of the kinematic provides visualization support for experiments on computer hand model. This software produces ground truth hand-pose vision-based hand pose estimation and tracking. The system data and corresponding images for use in computer vision allows users to save time in data generation and easily create experiments. It also produces ground truth data for the hand for any hand gestures. We have designed and implemented this user use in computer vision experiments [16]. The UI design and interface with the consideration of usability goals and software implementation for this simulation system followed accepted engineering issues. standards in UI design and software engineering. Keywords: GUI, User Interface Design, Virtual Reality, The organization of this paper is as follows: Section 2 briefly Software Engineering, HCI. -
An Empirical Study in Pen-Centric User Interfaces: Diagramming
EUROGRAPHICS Workshop on Sketch-Based Interfaces and Modeling (2008) C. Alvarado and M.- P. Cani (Editors) An Empirical Study in Pen-Centric User Interfaces: Diagramming Andrew S. Forsberg1, Andrew Bragdon1, Joseph J. LaViola Jr.2, Sashi Raghupathy3, Robert C. Zeleznik1 1Brown University, Providence, RI, USA 2University of Central Florida, Orlando, FL, USA 3Microsoft Corporation, Redmond, WA, USA Abstract We present a user study aimed at helping understand the applicability of pen-computing in desktop environments. The study applied three mouse-and-keyboard-based and three pen-based interaction techniques to six variations of a diagramming task. We ran 18 subjects from a general population and the key finding was that while the mouse and keyboard techniques generally were comparable or faster than the pen techniques, subjects ranked pen techniques higher and enjoyed them more. Our contribution is the results from a formal user study that suggests there is a broader applicability and subjective preference for pen user interfaces than the niche PDA and mobile market they currently serve. Categories and Subject Descriptors (according to ACM CCS): H.5.2 [User Interfaces]: Evaluation/Methodology 1. Introduction ficially appears pen-centric, users will in fact derive a sig- nificant benefit from using a pen-based interface. Our ap- Research on pen computing can be traced back at least to proach is to quantify formally, through head-to-head evalua- the early 60’s. Curiously though, there is little formal un- tion, user performance and relative preference for a represen- derstanding of when, where, and for whom pen comput- tative sampling of both keyboard and mouse, and pen-based ing is the user interface of choice. -
ACE-2019-Query-Builder-And-Tree
Copyright © 2019 by Aras Corporation. This material may be distributed only subject to the terms and conditions set forth in the Open Publication License, V1.0 or later (the latest version is presently available at http://www.opencontent.org/openpub/). Distribution of substantively modified versions of this document is prohibited without the explicit permission of the copyright holder. Distribution of the work or derivative of the work in any standard (paper) book form for a commercial purpose is prohibited unless prior permission is obtained from the copyright holder. Aras Innovator, Aras, and the Aras Corp "A" logo are registered trademarks of Aras Corporation in the United States and other countries. All other trademarks referenced herein are the property of their respective owners. Microsoft, Office, SQL Server, IIS and Windows are either registered trademarks or trademarks of Microsoft Corporation in the United States and/or other countries. Notice of Liability The information contained in this document is distributed on an "As Is" basis, without warranty of any kind, express or implied, including, but not limited to, the implied warranties of merchantability and fitness for a particular purpose or a warranty of non-infringement. Aras shall have no liability to any person or entity with respect to any loss or damage caused or alleged to be caused directly or indirectly by the information contained in this document or by the software or hardware products described herein. Copyright © 2019 by Aras Corporation. This material may be distributed only subject to the terms and conditions set forth in the Open Publication License, V1.0 or later (the latest version is presently available at http://www.opencontent.org/openpub/). -
Kurzweil 1000 Version 12 New Features
Kurzweil 1000 Version 12 New Features For the most up-to-date feature information, refer to the Readme file on the product CD. The following is a summary of what’s new in Version 12. For complete details, go to the online Manual by pressing Alt+H+O. Where applicable, Search key words are provided for you to use in the online Manual. • While you may not notice any difference, the internal structure of Kurzweil 1000 Version 12 has been overhauled and now uses Microsoft .NET Framework. The intent is to make it easier for Cambium Learning Technologies to develop features for the product going forward. • Note that Kurzweil 1000 Version 12 now supports 64-bit operating systems and Microsoft Windows 7 operating system. • As always, Kurzweil 1000 has the latest OCR engines, FineReader 9.0.1 and ScanSoft 16.2. The new ScanSoft version includes recognition languages from the Sami family. • An especially exciting new feature is the New User Wizard, a set of topics that introduces and walks new users through a number of Kurzweil 1000 features and preference setups. It appears when you start up Kurzweil 1000, but can be disabled and accessed from the Help menu by pressing Alt+H+W. (Search: New User Wizard.) • Currency Recognition has been updated to support new bills. Note that Currency Recognition now requires a color scanner. • New features and enhancements in reference tools include: 1. updates of the American Heritage Dictionary and Roget’s Thesaurus. 2. the ability to find up to 114 of your previously looked up entries; and last but not least the addition to dictionary and thesaurus lookup of human pronunciations and Anagrams. -
Pen Interfaces
Understanding the Pen Input Modality Presented at the Workshop on W3C MMI Architecture and Interfaces Nov 17, 2007 Sriganesh “Sri-G” Madhvanath Hewlett-Packard Labs, Bangalore, India [email protected] © 2006 Hewlett-Packard Development Company, L.P. The information contained herein is subject to change without notice Objective • Briefly describe different aspects of pen input • Provide some food for thought … Nov 17, 2007 Workshop on W3C MMI Architecture and Interfaces Unimodal input in the context of Multimodal Interfaces • Multimodal interfaces are frequently used unimodally − Based on • perceived suitability of modality to task • User experience, expertise and preference • It is important that a multimodal interface provide full support for individual modalities − “Multimodality” cannot be a substitute for incomplete/immature support for individual modalities Nov 17, 2007 Workshop on W3C MMI Architecture and Interfaces Pen Computing • Very long history … predates most other input modalities − Light pen was invented in 1957, mouse in 1963 ! • Several well-studied aspects: − Hardware − Interface − Handwriting recognition − Applications • Many famous failures (Go, Newton, CrossPad) • Enjoying resurgence since 90s because of PDAs and TabletPCs − New technologies such as Digital Paper (e.g. Anoto) and Touch allow more natural and “wow” experiences Nov 17, 2007 Workshop on W3C MMI Architecture and Interfaces Pen/Digitizer Hardware … • Objective: Detect pen position, maybe more • Various technologies with own limitations and characteristics (and new ones still being developed !) − Passive stylus • Touchscreens on PDAs, some tablets • Capacitive touchpads on laptops (Synaptics) • Vision techniques • IR sensors in bezel (NextWindow) − Active stylus • IR + ultrasonic (Pegasus, Mimeo) • Electromagnetic (Wacom) • Camera in pen tip & dots on paper (Anoto) • Wide variation in form − Scale: mobile phone to whiteboard (e.g. -
Veyon User Manual Release 4.1.91
Veyon User Manual Release 4.1.91 Veyon Community Mar 21, 2019 Contents 1 Introduction 1 1.1 Program start and login.........................................1 1.2 User interface...............................................2 1.3 Status bar.................................................2 1.4 Toolbar..................................................3 1.5 Computer select panel..........................................3 1.6 Screenshots panel............................................4 2 Program features 7 2.1 Using functions on individual computers................................7 2.2 Monitoring mode.............................................8 2.3 Demonstration mode...........................................8 2.4 Lock screens...............................................9 2.5 Remote access..............................................9 2.6 Power on, restart and shutdown computers............................... 11 2.7 Log off users............................................... 12 2.8 Send text message............................................ 12 2.9 Run program............................................... 13 2.10 Open website............................................... 13 2.11 Screenshot................................................ 14 3 FAQ - Frequently Asked Questions 15 3.1 Can other users see my screen?..................................... 15 3.2 How frequently are the computer thumbnails updated?......................... 15 3.3 What happens if I accidentally close the Veyon Master application window?.............. 15 3.4