The Phone with the Flow: Combining Touch + Optical Flow in Mobile Instruments Cagan Arslan, Florent Berthaut, Jean Martinet, Ioan Marius Bilasco, Laurent Grisoni

Total Page:16

File Type:pdf, Size:1020Kb

The Phone with the Flow: Combining Touch + Optical Flow in Mobile Instruments Cagan Arslan, Florent Berthaut, Jean Martinet, Ioan Marius Bilasco, Laurent Grisoni The Phone with the Flow: Combining Touch + Optical Flow in Mobile Instruments Cagan Arslan, Florent Berthaut, Jean Martinet, Ioan Marius Bilasco, Laurent Grisoni To cite this version: Cagan Arslan, Florent Berthaut, Jean Martinet, Ioan Marius Bilasco, Laurent Grisoni. The Phone with the Flow: Combining Touch + Optical Flow in Mobile Instruments. NIME 2018 - 18th Interna- tional Conference on New Interfaces for Musical Expression, Jun 2018, Blacksburg, VA, United States. hal-01809210 HAL Id: hal-01809210 https://hal.archives-ouvertes.fr/hal-01809210 Submitted on 6 Jun 2018 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. The Phone with the Flow: Combining Touch + Optical Flow in Mobile Instruments Cagan Arslan, Florent Berthaut, Jean Martinet, Ioan Marius Bilasco, Laurent Grisoni CRIStAL, CNRS, University of Lille, France {cagan.arslan,florent.berthaut,jean.martinet,marius.bilasco,laurent.grisoni}@univ-lille.fr ABSTRACT to gather position and motion information that are mapped Mobile devices have been a promising platform for musical to MIDI controls [15] [14]. performance thanks to the various sensors readily available However, to our knowledge, built-in cameras of mobile on board. In particular, mobile cameras can provide rich devices have not been used for sonification of moving el- input as they can capture a wide variety of user gestures ements in the camera image. Controlling sound through or environment dynamics. However, this raw camera input movement has always been of interest and can be traced only provides continuous parameters and requires expensive back to Kurenniemi's DIMI-O [12] and Rokeby's Very Ner- computation. In this paper, we propose combining camera vous System [1]. More recent examples such as [13][7][8] [4] based motion/gesture input with the touch input, in order use difference between images and optical flow to represent to filter movement information both temporally and spa- the moving parts of the image, but the methods have been tially, thus increasing expressiveness while reducing compu- either too simple to extract rich information or too heavy tation time. We present a design space which demonstrates to be run on mobile devices. The details of optical flow the diversity of interactions that our technique enables. We methods will be further discussed in section 2.2. also report the results of a user study in which we observe In this paper, we propose combining visual movement de- how musicians appropriate the interaction space with an tection with the touchscreen of mobile devices in order to example instrument. reduce the computation time and to open expressive op- portunities. We analyze these possibilities using a design space. We also study how users appropriate the instrument Author Keywords for two different movement sources: the user's gestures and Mobile instruments, Optical flow, touchscreen the environment. CCS Concepts 2. THE PHONE WITH THE FLOW In this section, we first present the main approach behind •Applied computing ! Sound and music computing; our system. It is currently implemented as an Android App •Computing methodologies ! Motion capture; (www.caganarslan.info/pwf.html). Optical flow features, extracted as described in the next section, are then mapped 1. INTRODUCTION to sound synthesis parameters. The synthesis can be done Many sensing capabilities of mobile devices have been ex- either directly on the mobile device, with restrictions on the plored for musical interaction [3]. Most of the mobile de- complexity of the synthesis due to limited computing capa- vices today includes various sensors such as a microphone, bilities of the mobile device, or the features can be sent to motion sensors, a touch screen, and multiple cameras. Hence, external musical software via OpenSoundControl messages. numerous musical interaction devices were designed around We also describe a design space of the interaction and smartphones and tablets [5] [2] [17], in particular using mappings opportunities afforded by our system. the discrete and continuous input capabilities of the touch- 2.1 From the Scene to the Touchscreen screen. Among these sensors however, the use of built-in camera of mobile devices has been little explored. The sim- Our system relies on real-time spatial and temporal filtering plest approach is to use the camera as a tone-hole. Ananya of rich motion information provided by the built-in cameras et al. [2] use the average gray-scale value of the camera of mobile devices. Fig. 1 depicts an example scenario of use input image to detect if the lens is covered. Covering or of The Phone with the Flow. uncovering of the lens modifies the pitch of the sound simi- Movement-rich Environment: The physical scene cap- larly to a tone-hole. Keefe and Essl [9] used low-level visual tured by the built-in camera offers a large range of move- features of the input image such as the edginess to create ments, with various periodicity, directions and sources. The mobile music performances with visual contributions. Cam- sources can be artificial (displays, mechanisms), natural era image has also been used to track visual reference points (weather, animals, plants), or originate from people (user's body, other musicians, spectators, by-standers). Choosing the Interaction Space Mobile cameras have a limited field of view but their portability enables explo- ration of the surroundings by a point-and-shoot approach. Unlike fixed installations, the mobile camera's field of view Licensed under a Creative Commons Attribution can be changed without effort by simply changing its posi- 4.0 International License (CC BY 4.0). Copyright tion and orientation. When the camera aims at a part of remains with the author(s). the movement-rich environment, objects in the field of view NIME’18, June 3-6, 2018, Blacksburg, Virginia, USA. are captured, providing the user with a visual feedback of (a) A person, a computer screen, a (b) The user chooses a movement (c) The user selects a region on the fan, objects passing by the window are source screen movement sources. Figure 1: From scene to the the touchscreen. their movement sources. The user is then free to interact a pyramidal Lucas-Kanade method[18] to track the sparsely in the combined interaction volume [11] of the camera and identified points. He also discusses the potential mappings the touchscreen. of the flow field to sound parameters. However he concludes Filtering the Movements Once the movement sources that the flow fields' temporal resolution is poor and the fea- are displayed on the touchscreen, the user focuses on a re- ture detection is not robust. CaMuS2[14] estimates the gion of interest by touching it. The touch selects a region optical flow from a camera phone. A 176x144 pixels image of the image to be analyzed further for detailed movement is divided into 22x18 blocks and cross-correlations between information. The user can change the position of the region successive pairs of block images are calculated to obtain by dragging their finger on the screen, alter its size, com- the optical flow. Because only 4 points are sampled in each bine multiple regions and switch between them. The touch block, the algorithm performs quickly. The simplicity of the input thus enables filtering of the movements. method allows obtaining the global translation and rotation at 15fps, but the system is unable to provide rich motion 2.2 Optical Flow information from the moving objects in the image. Optical flow is the distribution of apparent velocities of In recent years, there have been advances both in smart- moving brightness patterns in an image [6]. The movement phone hardware (improved cameras, CPU and GPU) and indicates a motion resulting from the relation between cap- optical flow algorithms, which open up new possibilities for tured objects and the viewer. The motion at a point can movement detection in real-time. be represented with ∆x and ∆y, the horizontal and the In 2016, Kroeger et al. [10] introduced DIS-Flow, an op- vertical displacement respectively. Therefore displacement tical flow estimation method that is one order of magnitude vector can be expressed as v(∆x; ∆y), but also in polar faster than others and produces roughly similar flow quality. coordinates v(r; Θ) to represent the amplitude and the di- Their objective is to trade-off a less accurate flow estima- rection of the movement for a pixel. Fig. 2 shows a color tion for large decreases in run-time for time critical tasks. coded representation of the optical flow. This method provides a dense flow output, in which ev- ery pixel is associated to a displacement vector v(∆x; ∆y). This advancement enables us to envision the creation of an innovative design space that allows richer interaction and sonification methods. Optical flow computation is noisy and presents important motion discontinuities as it measures the displacement in the 2D image plane of the real 3D world. Filtering out noise and focusing on the coherent motion information is the key to successful optical flow interaction. Our approach to extracting features from raw flow data is as follows. First of all, for every region of interest, the pixels that have a Figure 2: Color representation of the optical flow. (Top) smaller displacement value than a threshold are discarded Color space: hue indicates direction, saturation indicates to eliminate the noise (2). We identified key information to amplitude.
Recommended publications
  • No Slide Title
    December 2007 Updated 12/17/07 AT&T Mobility December ~ Washington Government WSCA – AT&T Mobility Device and Rate Plan Update for 2007 All Offers for Government Use ONLY On the attached pages you will find updates to the December WSCA pricing for Equipment and Rate Plans. Table of Contents Page 2: Voice Rate Plans Page 3: FREE Cellular Phones Page 4: Cellular Phones Page 5: BlackBerry Devices Page 6: SmartPhones (Microsoft, Nokia, Palm) Page 7: Push-to-Talk Devices Page 8: Aircards Page 9: Accessories Page 10: Nationwide Coverage Map December Highlights: ¾New Blackberry 8310 Curve! ~ Onboard GPS ~ Available in Titanium & Red! ¾ New Samsung Blackjack 2 – GPS ~ Available in Black & Red! ¾ Air Card Promotion: Sierra AC881 & Option GT Max FREE! *Certain devices may not be shown due to policy or otherwise WSCA Pricing. For more Information Contact: Prices and Promotions subject to change without notice Rob Holden All Offers for Government Use ONLY 425-580-7741 Master Price Agreement: T07-MST-069 [email protected] All plans receive WSCA 20% discount on monthly recurring service charges December 2007 December 2007 Updated 12/17/07 AT&T Mobility Oregon Government WSCA All plans receive 20% additional discount off of monthly recurring charges! AT&T Mobility Calling Plans REGIONAL Plan NATION Plans (Free Roaming and Long Distance Nationwide) Monthly Fee $9.99 (Rate Code ODNBRDS11) $39.99 $59.99 $79.99 $99.99 $149.99 Included mins 0 450 900 1,350 2,000 4,000 5000 N & W, Unlimited Nights & Weekends, Unlimited Mobile to 1000 Mobile to Mobile Unlim
    [Show full text]
  • Down and Dirty Camera Tricks Even for Your CELL PHONE
    Down and Dirty Camera Tricks Even for Your CELL PHONE Educational Seminar GCSAA Presenter - John R. Johnson Creating Images That Count The often quoted Yogi said, “When you come to a Fork in the Road – Take it!” What he meant was . “When you have a Decision in Life – Make it.” Let’s Make Good Decisions with Our Cameras I Communicate With Images You’ve known me for Years. Good Photography Is How You Can Communicate This one is from the Media Moses Pointe – WA Which Is Better? Same Course, 100 yards away – Shot by a Pro Moses Pointe – WA Ten Tricks That Work . On Big Cameras Too. Note image to left is cell phone shot of me shooting in NM See, Even Pros Use Cell Phones So Let’s Get Started. I Have My EYE On YOU Photography Must-Haves Light Exposure Composition This is a Cell Phone Photograph #1 - Spectacular Light = Spectacular Photography Colbert Hills - KS Light from the side or slightly behind. Cell phones require tons of light, so be sure it is BRIGHT. Sunsets can’t hurt either. #2 – Make it Interesting Change your angle, go higher, go lower, look for the unusual. Resist the temptation to just stand and shoot. This is a Cell Phone Photograph Mt. Rainier Coming Home #2 – Make it Interesting Same trip, but I shot it from Space just before the Shuttle re-entry . OK, just kidding, but this is a real shot, on a flight so experiment and expand your vision. This is a Cell Phone Photograph #3 – Get Closer In This Case Lower Too Typically, the lens is wide angle, so things are too small, when you try to enlarge, they get blurry, so get closer to start.
    [Show full text]
  • Visual Mobile Communication Camera Phone Photo Messages As Ritual Communication and Mediated Presence Mikko Villi’S Background Is in Com- Munication Studies
    mikko villi Visual mobile communication Camera phone photo messages as ritual communication and mediated presence Mikko Villi’s background is in com- munication studies. He has worked as a researcher at the University of Art and Design Helsinki, University of Tampere and University of Helsin- ki, where he has also held the position of university lecturer. Currently, he works as coordinator of educational operations at Aalto University. Villi has both researched and taught sub- jects related to mobile communica- tion, visual communication, social media, multi-channel publishing and media convergence. Visual mobile communication mikko villi Visual mobile communication Camera phone photo messages as ritual communication and mediated presence Aalto University School of Art and Design Publication series A 103 www.taik.fi/bookshop © Mikko Villi Graphic design: Laura Villi Original cover photo: Ville Karppanen !"#$ 978-952-60-0006-0 !""$ 0782-1832 Cover board: Ensocoat 250, g/m2 Paper: Munken Lynx 1.13, 120 g/m2 Typography: Chronicle Text G1 and Chronicle Text G2 Printed at %" Bookwell Ltd. Finland, Jyväskylä, 2010 Contents acknowledgements 9 1 introduction 13 2 Frame of research 17 2.1 Theoretical framework 19 2.1.1 Ritual communication 19 2.1.2 Mediated presence 21 2.2 Motivation and contribution of the study 24 2.3 Mobile communication as a field of research 27 2.4 Literature on photo messaging 30 2.5 The Arcada study 40 2.6 Outline of the study 45 3 Camera phone photography 49 3.1 The real camera and the spare camera 51 3.2 Networked photography
    [Show full text]
  • Of-Device Finger Gestures
    LensGesture: Augmenting Mobile Interactions with Back- of-Device Finger Gestures Xiang Xiao♦, Teng Han§, Jingtao Wang♦ ♦Department of Computer Science §Intelligent Systems Program University of Pittsburgh University of Pittsburgh 210 S Bouquet Street 210 S Bouquet Street Pittsburgh, PA 15260, USA Pittsburgh, PA 15260, USA {xiangxiao, jingtaow}@cs.pitt.edu [email protected] ABSTRACT becomes the only channel of input for mobile devices, leading to We present LensGesture, a pure software approach for the notorious "fat finger problem" [2, 22], the “occlusion augmenting mobile interactions with back-of-device finger problem” [2, 18], and the "reachability problem" [20]. In gestures. LensGesture detects full and partial occlusion as well as contrast, the more responsive, precise index finger remains idle on the dynamic swiping of fingers on the camera lens by analyzing the back of mobile devices throughout the interactions. Because image sequences captured by the built-in camera in real time. We of this, many compelling techniques for mobile devices, such as report the feasibility and implementation of LensGesture as well multi-touch, became challenging to perform in such a "situational as newly supported interactions. Through offline benchmarking impairment" [14] setting. and a 16-subject user study, we found that 1) LensGesture is easy Many new techniques have been proposed to address these to learn, intuitive to use, and can serve as an effective challenges, from adding new hardware [2, 15, 18, 19] and new supplemental input channel for today's smartphones; 2) input modality, to changing the default behavior of applications LensGesture can be detected reliably in real time; 3) LensGesture for certain tasks [22].
    [Show full text]
  • A Guide to Smartphone Astrophotography National Aeronautics and Space Administration
    National Aeronautics and Space Administration A Guide to Smartphone Astrophotography National Aeronautics and Space Administration A Guide to Smartphone Astrophotography A Guide to Smartphone Astrophotography Dr. Sten Odenwald NASA Space Science Education Consortium Goddard Space Flight Center Greenbelt, Maryland Cover designs and editing by Abbey Interrante Cover illustrations Front: Aurora (Elizabeth Macdonald), moon (Spencer Collins), star trails (Donald Noor), Orion nebula (Christian Harris), solar eclipse (Christopher Jones), Milky Way (Shun-Chia Yang), satellite streaks (Stanislav Kaniansky),sunspot (Michael Seeboerger-Weichselbaum),sun dogs (Billy Heather). Back: Milky Way (Gabriel Clark) Two front cover designs are provided with this book. To conserve toner, begin document printing with the second cover. This product is supported by NASA under cooperative agreement number NNH15ZDA004C. [1] Table of Contents Introduction.................................................................................................................................................... 5 How to use this book ..................................................................................................................................... 9 1.0 Light Pollution ....................................................................................................................................... 12 2.0 Cameras ................................................................................................................................................
    [Show full text]
  • Smartwatch Instructions
    SMARTWATCH INSTRUCTIONS IOS To download our app, search for RADLEY LONDON WATCH app or scan the appropriate QR code for installation. Use Thank you for purchasing your your smart phone camera or QR code RADLEY LONDON smartwatch reader app. Please follow these instructions The RADLEY LONDON WATCH app will carefully and we’ll soon have you ANDROID work on the following smartphones: up & running with your new watch. Android (version 4.4 and above) Apple IOS (version 8.0 and above) Please ensure the app is the latest version and have Bluetooth 5.0 QUICK START GUIDE CONNECTING BAND TO MOBILE PHONE CHARGING GUIDE 1. Enable Bluetooth on your smartphone. Charging for automatic power-on: 2. Open RADLEY App. on your smart phone, set your profile within settings. 1. Line up the 2 pins to the charging spots on the back of the watch case. 3. Tap connect device within settings. 2. The pins are magnetic and will secure the charger in place. 4. Select Smart Bracelet as shown. 3. Plug in the USB cable to a power source. Once connected, a charging icon 5. Follow the steps, search for RADLEY LONDON device will appear on screen. name and connect. 4. Please charge the band when it displays low battery. 6. Within the notifications section, select the 5. A 2 hour charge time is sufficient to fully charge your smartwatch. Please notification types you want your band to receive. ensure you do not exceed a 2 hour charge period. OPERATION INTRODUCTION DISCONNECTING YOUR SMART WATCH & PHONE • To switch on the watch, press and hold the side button until the display is on.
    [Show full text]
  • Optical Design of the 13 Mega-Pixels Mobile Phone Camera Pengbo Chen
    3rd International Conference on Materials Engineering, Manufacturing Technology and Control (ICMEMTC 2016) Optical design of the 13 mega-pixels mobile phone camera Pengbo Chen,Xingyu Gaoa,* Institute of Intelligent Opto-mechatronics and Advanced Manufacture Technology, College of Electromechanical Engineering, Guilin University of Electronic Technology, Guilin 541004,China. a,*[email protected] Corresponding Author:Xingyu Gao Keywords: Optical design ; Mobile phone lens; Aspherical surface; Abstract. According to the market demand for high resolution cell phone camera,a 13 mega-pixels camera lens applying aspherical surfaces is designed using the ZEMAX optical design software. The lens is consisted by four pieces of aspherical plastic lens, an optical filter and a piece of protection glass. The lens has the optical performance of F-number 2.8, viewing angle 76°, focal length 4.4 mm, back focal length 0.58 mm, and total length 5.6 mm. After the tolerance analyzing, its tolerance meets the processing requirements, which can satisfy the requirements of the commercial sell phone. 1. Introduction Since 2000, Japan's Sharp Co launched the world's first camera phone. Camera function has become an essential feature of the phone. In 2003, Sony launched the first mobile phone with camera function, model for Ericsson T618, this phone installed 10 million pixel mobile phone lens, which is the earliest a camera phone, marking China's entry into the era of mobile phone camera on the Chinese market. In 2006, Zhu Rihong [1], who designed a f / 3.2, field of view angle of 55 degrees, pixel reached 160 million of the lens, the lens system as the total length of 5mm.In 2008, Liu Maochao [2], who designed a F/2.85, field of view angle of 62 degrees, pixel to 3 million of the lens, the system length of the lens is 5.26mm.In 2009, Li Wenjing [3], who designed a F/2.8, the field of view angle of 65 degrees, pixel 5 million of the mobile phone lens, the total length of the lens system is 5.8mm.
    [Show full text]
  • Samsung Galaxy Camera Forensics
    Combining technical and legal expertise to deliver investigative, discovery and forensic solutions worldwide. SAMSUNG GALAXY April 11, 2013 Introduction. The Samsung Galaxy camera was Operating System: Abstract released on November 16, 2012. This Android 4.1 (Jellybean) Samsung Galaxy Camera device has the potential to replace mobile Network:2G, 3G or 4G (LTE) phones, as it has the same functionality GSM, HSPA+ Forensics of a smartphone, with the additional perk of a high quality camera. This creates an Processor:1.4 GHz Quad Core attractive incentive to buy the camera, Memory:microSD, 4GB on The purpose of this project which could lead to the possibility of it board, 1GB RAM was to determine whether or not forensics on the Samsung Galaxy becoming more popular. Connectivity: WiFI As the camera’s popularity rises in the 802.11a/b/g/n, WiFi hotspot camera was possible. Although market, and more users purchase the the camera runs an Android device, the risk of the camera being used Bluetooth:Yes operating system, there was still a chance that no data could be in an illicit manner rises as well. As the GPS:Yes extracted, as forensics on this Samsung Galaxy camera is now a part device had never been done of an investigator’s scope, understanding Table 1. before. To begin the process of where any evidence can be retrieved Goals. this project, as much data as is crucial. Using several different The goal for this project was to possible had to be created on forensic tools, any data that could be of develop an informational guide for the the camera by utilizing all of the applications and features that evidentiary value is detailed in this paper.
    [Show full text]
  • Sony Ericsson C905a 8.1 Megapixel Cyber-Shot™ Camera Phone
    Sony Ericsson C905a 8.1 Megapixel Cyber-shot™ Camera Phone The Sony Ericsson phone designed for the photographer in you. Picture this: A Sony Ericsson phone that’s also an 8.1 megapixel Cyber-shot™ camera. With the C905a, it’s a snap to create brilliant images with user-friendly features such as auto-focus, face detection, and xenon flash with red-eye reduction. Smart Contrast helps to brighten underexposed subjects. BestPic™ captures subjects that just won’t sit still. Scene options provide a simple method to set camera settings automatically for getting 8.1 Megapixel Cyber-shot™ Camera the best exposure out of a particular situation. GPS with AT&T Navigator The C905a is ready when you are to capture the moments you won’t want to miss. All this in a AT&T Video Share stylish 3G phone that offers all the high quality AT&T Mobile Music services and performance you expect from AT&T. Email and Messaging For more information, visit att.com/wireless or call 866-MOBILITY. Screen images are simulated or enhanced. The Liquid Identity logo is a trademark or registered trademark of Sony Ericsson Mobile Communications AB. Sony and Cyber-shot are trademarks or registered trademarks of Sony Corporation. Ericsson is a trademark or registered trademark of Telefonaktiebolaget LM Ericsson. Yahoo! is a trademark or a registered trademark of Yahoo! Inc. Apple is a trademark or a registered trademark of Apple Computer, Inc. Facebook is a trademark or a registered trademark of Facebook, Inc. Wal-Mart is a trademark or registered trademark of Wal-Mart Stores Inc.
    [Show full text]
  • Autofocus Measurement for Imaging Devices
    Autofocus measurement for imaging devices Pierre Robisson, Jean-Benoit Jourdain, Wolf Hauser Clement´ Viard, Fred´ eric´ Guichard DxO Labs, 3 rue Nationale 92100 Boulogne-Billancourt FRANCE Abstract els of the captured image increases with correct focus [3], [2]. We propose an objective measurement protocol to evaluate One image at a single focus position is not sufficient for focus- the autofocus performance of a digital still camera. As most pic- ing with this technology. Instead, multiple images from differ- tures today are taken with smartphones, we have designed the first ent focus positions must be compared, adjusting the focus until implementation of this protocol for devices with touchscreen trig- the maximum contrast is detected [4], [5]. This technology has ger. The lab evaluation must match with the users’ real-world ex- three major inconveniences. First, the camera never can be sure perience. Users expect to have an autofocus that is both accurate whether it is in focus or not. To confirm that the focus is correct, and fast, so that every picture from their smartphone is sharp and it has to move the lens out of the right position and back. Sec- captured precisely when they press the shutter button. There is ond, the system does not know whether it should move the lens a strong need for an objective measurement to help users choose closer to or farther away from the sensor. It has to start moving the best device for their usage and to help camera manufacturers the lens, observe how contrast changes, and possibly switch di- quantify their performance and benchmark different technologies.
    [Show full text]
  • Digital Cameras and Expanding Mobile Handset Functions: Competing Value Chains in the Consumer Imaging Industry
    Digital Cameras and Expanding Mobile Handset Functions: Competing Value Chains in the Consumer Imaging Industry Yuri Park, Seoul National University, Techno-Economics and Policy Program, Seoul, South Korea Milton Mueller, Syracuse University, School of Information Studies, Syracuse, New York We selected six attributes for purchasing devices. (See Table Abstract 1) The alternatives that we considered in this study are confined to This paper explores the competition between specialized, digital cameras and camera phones in order to isolate the stand-alone digital cameras and digital cameras that are bundled relationship between them. with mobile telephones. It uses conjoint survey research to Our conjoint survey of capture devices relied on six attributes. determine whether camera phones and digital cameras are The first attribute was resolution; i.e., the quality of the pictures complements or substitutes, and to find out how much consumer taken by the device. The resolution of most camera phones is lower value specific features of cameras, such as LCD screen size, an than the norm for digital cameras, but can be expected to increase.1 internet connection, zoom, and so on. But no one knows the extent to which consumers want higher resolution. If consumers are satisfied with moderate resolution Introduction levels, there will be no reason to concentrate on developing camera Handheld phones now can provide digital cameras, PDAs, phones with significantly higher resolution. The second attribute of storage and playback of images, video and music. The convergence a capture device is the zoom function. This includes both optical of mobile handsets with digital imaging capabilities in particular has and digital zoom.
    [Show full text]
  • DXOMARK Objective Video Quality Measurements
    https://doi.org/10.2352/ISSN.2470-1173.2020.9.IQSP-166 © 2020, Society for Imaging Science and Technology DXOMARK Objective Video Quality Measurements Emilie Baudin1, Franc¸ois-Xavier Bucher2, Laurent Chanas1 and Fred´ eric´ Guichard1 1: DXOMARK, Boulogne-Billancourt, France 2: Apple, Cupertino, CA, USA Abstract thus they cannot be considered comprehensive video quality mea- Video capture is becoming more and more widespread. The surement protocols. Temporal aspects of video quality begin to be technical advances of consumer devices have led to improved studied; for instance, a perceptual study of auto-exposure adapta- video quality and to a variety of new use cases presented by so- tion has been presented by Oh et al. [8]. To the best of our knowl- cial media and artificial intelligence applications. Device man- edge, there is not as yet a measurement protocol with correlation ufacturers and users alike need to be able to compare different to subjective perception for video image quality. cameras. These devices may be smartphones, automotive com- The goal of this work is to provide a measurement protocol ponents, surveillance equipment, DSLRs, drones, action cameras, for the objective evaluation of video quality that reliably allows to etc. While quality standards and measurement protocols exist for compare different cameras. It should allow to automatically rank still images, there is still a need of measurement protocols for cameras in the same order a final user would do it. In this article, video quality. These need to include parts that are non-trivially we focus on three temporal metrics: exposure convergence, color adapted from photo protocols, particularly concerning the tem- convergence, and temporal visual noise.
    [Show full text]