FINAL PROGRAM Imai

Total Page:16

File Type:pdf, Size:1020Kb

FINAL PROGRAM Imai FINAL PROGRAM 14–18 February 2016 • San Francisco, CA, USA Photo: Diana Gonzalez. Photo: Stephen Keith. Photo: Diana Gonzalez. Photo: Francisco Imai. • SHORT COURSES • EXHIBITS • DEMONSTRATION SESSION • PLENARY TALKS • INTERACTIVE PAPER SESSION • SPECIAL EVENTS • TECHNICAL SESSIONS electronicimaging.org TO LOBBY 14–18 February 2016 Hilton San Francisco, Union Square San Francisco, California USA 2016 Symposium Chair Choon-Woo Kim Welcome Inha Univ. (Republic of Korea) On behalf of IS&T—the Society for Imaging Science and Technology—we would like to welcome you to the 28th annual International Symposium on Electronic Imaging. Imaging is pervasive in the human experience—from the way we view the world each day 2016 Symposium to the photographs we take on our smart phones to its exciting use in technologies related Co-Chair to national security, space exploration, entertainment, medical, and printing applications— Nitin Sampat and an increasingly vital part of our lives. Rochester Institute of Technology (USA) This week you have the opportunity to hear the latest research from the world’s leading ex- perts in imaging, image processing, sensors, color, and augmented/virtual reality/3D, to name but a few of the applications and technologies covered by the event. You also have many opportunities to develop both your career and business by networking with leading 2016 Short Course Chair researchers and entrepreneurs in the field. Majid Rabbani Eastman Kodak Co. The Electronic Imaging Symposium is the premier international meeting in this exciting (USA) technological area, one that brings together academic and industry colleagues to discuss topics on the forefront of research and innovation. We look forward to seeing you and welcoming you to this unique event. 2016 Short Course —Choon-Woo Kim and Nitin Sampat, EI2016 Symposium Co-chairs Co-Chair Mohamed-Chaker Larabi University of Poitiers (France) IS&T CORPORATE MEMBERS Sustaining Corporate Members / Symposium Sponsors ® Supporting Corporate Members BASF Corporation • FUJIFILM Corporation • Konica Minolta, Inc. • Lexmark International Inc. Donor Corporate Members Ball Packaging Europe Holding GmbH & Co KG • Cheran Digital Imaging & Consulting, Inc. Clariant Produkte (Deutschland) GmbH • Japanese Patent Office Kunshan Hisense Electronic Co. Ltd. • Ricoh Company, Ltd. Schoeller Technocell GmbH & Co KG • TREK, INC./TREK JAPAN KK IS&T Board of Directors July 2015 - June 2016 President: Vice Presidents UPCOMING IS&T EVENTS Geoff Woolfe, Sergio Goma, Canon Information Systems Qualcomm Technologies, Inc. April 19 – 22, 2016; Washington, DC Research Australia Pty. Ltd Archiving 2016 (CISRA) Teruaki Mitsuya, Ricoh Company, Ltd. September 12 – 16, 2016; Manchester, UK Executive Vice President Printing for Fabrication Steven Simske, Majid Rabbani, HP Inc. Eastman Kodak Company 32nd International Conference on Digital Printing Technologies (formerly NIP) Conference Vice President James Stasiak, Francisco Hideki Imai, HP Inc. September 15 – 16, 2016; Manchester, UK Canon USA, Inc. Technologies in Digital Photo Fulfillment Radka Tezaur, Publications Vice President Intel Corporation Susan Farnand, November 7 – 11, 2016; San Diego, CA Rochester Institute of Werner Zapka, 24th Color and Imaging Conference (CIC24) Technology XaarJet AB Jan. 29 – Feb. 2, 2017; SFO/Burlingame, California Secretary Immediate Past President Electronic Imaging 2017 Steve Korol, Alan Hodgson, Evolutionary Technology Alan Hodgson Consulting Ltd. Jan. 28 – Feb. 1, 2018; SFO/Burlingame, California Treasurer Chapter Directors Electronic Imaging 2018 Scott Silence, Rochester: Michel Molaire, Xerox Corporation Molecular Glasses Tokyo: Jun-ichi Hanna, Learn more at www.imaging.org/ist/conferences/. Tokyo Institute of Technology A complete list of imaging-related meetings is at www.imaging.org/ist/conferences/events.cfm IS&T Executive Director Suzanne E. Grinnan Table of Contents Symposium Leadership ..................................................................................................... 4 Symposium Overview ....................................................................................................... 5 Plenary Speakers ............................................................................................................. 7 Special Events ................................................................................................................ 8 Short Course Daily Schedule ............................................................................................. 9 Keynotes and Invited Talks............................................................................................... 10 Joint Sessions Planned for EI 2016 ................................................................................... 12 Conferences and submission information by topic Human Perception and Cognition for Emerging Technologies Human Vision and Electronic Imaging (HVEI) 2016 (Rogowitz, Pappas, and de Ridder) ............ 15 Image Capture Systems Digital Photography and Mobile Imaging XII (Roland, Tezaur, and Wüller) ............................... 21 Image Sensors and Imaging Systems 2016 (Widenhorn, Dupret, and Darmont) ....................... 25 Mobile Devices and Multimedia: Enabling Technologies, Algorithms, and Applications 2016 (Creutzburg and Akopian) .................................................................... 28 Image Reproduction and Material Appearance Color Imaging XXI: Displaying, Processing, Hardcopy, and Applications (Eschbach, Marcu, and Rizzi) .......................................................................................... 31 Measuring, Modeling, and Reproducing Material Appearance 2016 (Imai, Ortiz-Segovia, and Urban) .................................................................................... 36 Document Processing and Media Security Document Recognition and Retrieval XXIII (Lamiroy and Antani) .............................................. 39 Media Watermarking, Security, and Forensics 2016 (Alattar and Memon) .............................. 42 Image and Video Processing, Quality, and Systems Computational Imaging XIV (Bouman and Sauer) ................................................................ 45 Image Processing: Algorithms and Systems XIV (Egiazarian, Agaian, and Gotchev) ................... 49 Image Quality and System Performance XIII (Larabi and Jenkin) .............................................. 53 Visual Information Processing and Communication VII (Guleryuz, Said, and Stevenson) .............. 57 Virtual and Augmented Reality, 3D, and Stereoscopic Systems 3D Image Processing, Measurement (3DIPM), and Applications 2016 (Puech and Sitnik)........... 60 The Engineering Reality of Virtual Reality 2016 (Dolinsky and McDowall) ................................ 63 Stereoscopic Displays and Applications XXVII (Woods, Holliman, Favalora, and Kawai) ............ 66 Real-time Image and Video Image Processing: Machine Vision Applications IX (Lam, Ngan, and Niel) .............................. 70 Intelligent Robots and Computer Vision XXXIII: Algorithms and Techniques (Casasent and Röning) ......... 73 Web and Mobile Imaging and Visualization Imaging and Multimedia Analytics in a Web and Mobile World 2016 (Allebach, Fan, and Lin) ......... 75 Video Surveillance and Transportation Imaging Applications 2016 (Saber) .............................. 78 Visualization and Data Analysis 2016 (Kao, Wischgoll, and Zhang) ..................................... 80 Short Courses ................................................................................................................ 84 General Information ....................................................................................................... 93 About the Organizer ...................................................................................................... 93 Author Index .................................................................................................................. 97 Plan Now to Participate Join us for Electronic Imaging 2017 Photo: Stephen Keith. Burlingame, California, 29 January – 2 February, 2017 electronicimaging.org 3 EI Symposium Leadership EI2016 Symposium Committee Symposium Chair Zhigang Fan, Apple Inc. Choon-Woo Kim, Inha University (Republic Gregg E. Favalora, VisionScope of Korea) Technologies, LLC Onur Guleryuz, LG Electronics Symposium Co-Chair MobileComm USA, Inc. Nitin Sampat, Rochester Institute of Atanas Gotchev, Tampere Univ. of Technology Technology (USA) Nicolas S. Holliman, Newcastle Univ. Francisco Imai, Canon USA, Inc. Short Course Chair Robin Jenkin, ON Semiconductor Corp. Majid Rabbani, Eastman Kodak Co. (USA) David Kao, NASA Ames Research Center Takashi Kawai, Waseda Univ. Short Course Co-Chair Edmund Y. Lam, The Univ. of Hong Kong Mohamed-Chaker Larabi, University of Bart Lamiroy, Univ. de Lorraine Poitiers (France) Mohamed-Chaker (Chaker) Larabi, Univ. of Poitiers At-large Conference Chair Representative Qian Lin, Hewlett-Packard Lab. Adnan Alattar, Digimarc Corp. (USA) Gabriel Marcu, Apple Inc. Ian McDowall, Fakespace Labs, Inc. Local Liaison Chair Nasir D. Memon, Polytechnic Institute of Photo: Can Balcioglu. Joyce Farrell, Stanford University (USA) New York Univ. Henry Ngan, Hong Kong Baptist Univ. Exhibit and Sponsorship Chair Kurt S. Niel, Upper Austria Univ. of Applied Kevin Matherson, Microsoft Corp. (USA) Sciences Maria V. Ortiz-Segovia, Océ Print Logic Past Symposium Chair Technologies Sheila Hemami, Northeastern University (USA) Thrasyvoulos
Recommended publications
  • RV - RA - RM Realidades Virtuales, Aumentadas Y Mixtas Índice General
    RV - RA - RM Realidades virtuales, aumentadas y mixtas Índice general 1 Definiciones 1 1.1 Realidad virtual ............................................ 1 1.1.1 Virtualidad .......................................... 1 1.1.2 Relación realidad/irrealidad ................................. 1 1.1.3 Inmersión y navegación .................................... 2 1.1.4 Usos ............................................. 2 1.1.5 Productos ........................................... 3 1.1.6 Técnicas de realidad virtual ................................. 6 1.1.7 Problemas de la realidad virtual ............................... 7 1.1.8 Tecnoética realidad virtual .................................. 8 1.1.9 Véase también ........................................ 9 1.1.10 Referencias .......................................... 9 1.1.11 Enlaces externos ....................................... 10 1.1.12 Bibliografía .......................................... 10 1.2 Realidad aumentada .......................................... 10 1.2.1 Definiciones ......................................... 11 1.2.2 Cronología .......................................... 11 1.2.3 Tecnología .......................................... 12 1.2.4 Técnicas de visualización ................................... 13 1.2.5 Elementos de la realidad aumentada ............................. 13 1.2.6 Niveles ............................................ 14 1.2.7 Aplicaciones ......................................... 14 1.2.8 Aplicaciones futuras ..................................... 16 1.2.9 Literatura
    [Show full text]
  • Augmented Reality Applied Tolanguage Translation
    Ana Rita de Tróia Salvado Licenciado em Ciências da Engenharia Electrotécnica e de Computadores Augmented Reality Applied to Language Translation Dissertação para obtenção do Grau de Mestre em Engenharia Electrotécnica e de Computadores Orientador : Prof. Dr. José António Barata de Oliveira, Prof. Auxiliar, Universidade Nova de Lisboa Júri: Presidente: Doutor João Paulo Branquinho Pimentão, FCT/UNL Arguente: Doutor Tiago Oliveira Machado de Figueiredo Cardoso, FCT/UNL Vogal: Doutor José António Barata de Oliveira, FCT/UNL September, 2015 iii Augmented Reality Applied to Language Translation Copyright c Ana Rita de Tróia Salvado, Faculdade de Ciências e Tecnologia, Universi- dade Nova de Lisboa A Faculdade de Ciências e Tecnologia e a Universidade Nova de Lisboa têm o direito, perpétuo e sem limites geográficos, de arquivar e publicar esta dissertação através de ex- emplares impressos reproduzidos em papel ou de forma digital, ou por qualquer outro meio conhecido ou que venha a ser inventado, e de a divulgar através de repositórios científicos e de admitir a sua cópia e distribuição com objectivos educacionais ou de in- vestigação, não comerciais, desde que seja dado crédito ao autor e editor. iv To my beloved family... vi Acknowledgements "Coming together is a beginning; keeping together is progress; working together is success." - Henry Ford. Life can only be truly enjoyed when people get together to create and share moments and memories. Greatness can be easily achieved by working together and being sup- ported by others. For this reason, I would like to save a special place in this work to thank people who were there and supported me during all this learning process.
    [Show full text]
  • Echtzeitübersetzung Dank Augmented Reality Und Spracherkennung Eine Untersuchung Am Beispiel Der Google Translate-App
    ECHTZEITÜBERSETZUNG DANK AUGMENTED REALITY UND SPRACHERKENNUNG EINE UNTERSUCHUNG AM BEISPIEL DER GOOGLE TRANSLATE-APP Julia Herb, BA Matrikelnummer: 01216413 MASTERARBEIT eingereicht im Rahmen des Masterstudiums Translationswissenschaft Spezialisierung: Fachkommunikation an der Leopold-Franzens-Universität Innsbruck Philologisch-Kulturwissenschaftliche Fakultät Institut für Translationswissenschaft betreut von: ao. Univ.-Prof. Mag. Dr. Peter Sandrini Innsbruck, am 22.07.2019 1 Inhaltsverzeichnis Abstract ........................................................................................................................................................... 1 Abbildungsverzeichnis .................................................................................................................................... 2 1. Aktualität des Themas ............................................................................................................................. 4 2. Augmented Reality: die neue Form der Multimedialität ........................................................................ 7 a. Definition und Abgrenzung ................................................................................................................ 7 b. Anwendungsbereiche der AR-Technologie ...................................................................................... 11 3. Verbmobil: neue Dimensionen der maschinellen Übersetzung dank Spracherkennung ....................... 13 a. Besonderheiten eines Echtzeit-Übersetzungs-Projekts ....................................................................
    [Show full text]
  • Machine Learning Is Going Mobile
    Machine learning is going mobile By David Schatsky ACHINE learning—the process by which Signals Mcomputers can get better at perform- ing tasks through exposure to data, rather • Google has introduced language translation than through explicit programming—requires software, using small neural networks opti- massive computational power, the kind usu- mized for mobile phones, which can per- ally found in clusters of energy-guzzling, form well without an Internet connection.1 cloud-based computer servers outfitted with specialized processors. But an emerging trend • Lenovo announced a mobile phone that promises to bring the power of machine learn- uses multiple sensors, high-speed image ing to mobile devices that may lack or have processing hardware, and specialized only intermittent online connectivity. This will Google software to support capabili- give rise to machines that sense, perceive, learn ties such as indoor wayfinding, precision from, and respond to their environment and measuring, and augmented reality even their users, enabling the emergence of new when offline.2 product categories, reshaping how businesses engage with customers, and transforming how • NVIDIA, a maker of graphics processing work gets done across industries. technology, introduced an embeddable module for computer vision applications Machine learning is going mobile in devices such as drones and autono- systems run in the cloud on powerful serv- mous vehicles that the company says ers, processing data such as digitized voice or consumes one-tenth the power of a photos that users upload. competing offering.3 Until recently, a typical smartphone lacked the power to perform such tasks without con- • Qualcomm introduced a new processor and necting to the cloud, except in limited ways.
    [Show full text]
  • Augmented Reality Based Word Translator Tabish Khan, Rishisingh Hora, Ashwin Bendre, Prof
    International Journal of Innovative Research in Computer Science & Technology (IJIRCST) ISSN: 2347-5552, Volume-2, Issue-2, March 2014 Augmented Reality Based Word Translator Tabish Khan, Rishisingh Hora, Ashwin Bendre, Prof. Sneha Tirth developed for iOS. The translation it provides is not always Abstract— People travel to different places not knowing 100 percent accurate. It has a hit rate of 90-95 percent. the language used in that region. Hence there is a need to translate these unknown words to recognizable text. This application is developed to help travelers who can get the translated text as an output of our application. Our application is able to recognize the text captured by a mobile phone camera, translate the text and display the translation result back onto the screen of the mobile phone in an augmented manner. Index Terms— Android, Augmented Reality, Optical Character Recognition, Word Translator. NTRODUCTION I. I The motivation of our application is to help tourist FIG 1: Output of word lens navigate in a foreign language environment. The The above figure depicts an example of the word lens application we developed enables the users to get text application which translates the words “Bienvenido Al translated as easy as a button click. The camera captures the Futuro” to the desired English language, the output being, text and returns the translated result in the real time. “Welcome to the future”. The system we developed includes automatic text detection, OCR and text translation. Although the current version of our application is limited to translation of a few languages, III. PROPOSED SYSTEM we can add a feature of text-to-speech, as an extension in Inspired by the existing application, we will use the same our application.
    [Show full text]
  • Final Thesis1
    Chapter 3. HISTORICAL BACKGROUND OF QR CODE & AR APPLICATION 3.1. What is a QR Code QR code (abbreviated from Quick Response Code) is the trademark for a type of matrix barcode (or two-dimensional bar code) A QR code is a matrix barcode readable by smart phones and mobile phones with cameras. They are sometimes referred to as 2d codes, 2d barcodes, or mobile codes. QR codes can hold much more information than a regular barcode. The information encoded in a QR code can be a URL, a phone number, an SMS message, a V-card, or any text. They are referred to as QR because they allow the contents to be decoded at high speed. QR codes were developed in 1994 by Denso-Wave, a Toyota subsidiary In Catalog records to offer patrons basic information about location of books and call number, users can scan the code and head to the stacks rather than writing or printing, this means, for example, a student who scans QR codes from a library OPAC in the middle of the night in his or her dorm room could head to the library stacks the next day, click open the app to refer to the call numbers of the titles scanned the night before, and quickly find the books. Also, expect to see QR code scanner/reader apps that allow users to tag, share, comment, collaborate, and more in the future. Taped to video/DVD cases, linking to mobile-friendly video trailers, Codes in the library stacks/end caps or magazine/journal areas that point to online electronic holdings of print materials or related subject guides, linking to library audio tours for orientations etc.
    [Show full text]
  • Arxiv:1611.08903V1 [Cs.LG] 27 Nov 2016 20 Percent Increases in Cash Collections [20]
    Should I use TensorFlow? An evaluation of TensorFlow and its potential to replace pure Python implementations in Machine Learning Martin Schrimpf1;2;3 1 Augsburg University 2 Technische Universit¨atM¨unchen 3 Ludwig-Maximilians-Universit¨atM¨unchen Seminar \Human-Machine Interaction and Machine Learning" Supervisor: Elisabeth Andr´e Advisor: Dominik Schiller Abstract. Google's Machine Learning framework TensorFlow was open- sourced in November 2015 [1] and has since built a growing community around it. TensorFlow is supposed to be flexible for research purposes while also allowing its models to be deployed productively [7]. This work is aimed towards people with experience in Machine Learning consider- ing whether they should use TensorFlow in their environment. Several aspects of the framework important for such a decision are examined, such as the heterogenity, extensibility and its computation graph. A pure Python implementation of linear classification is compared with an im- plementation utilizing TensorFlow. I also contrast TensorFlow to other popular frameworks with respect to modeling capability, deployment and performance and give a brief description of the current adaption of the framework. 1 Introduction The rapidly growing field of Machine Learning has been gaining more and more attention, both in academia and in businesses that have realized the added value. For instance, according to a McKinsey report, more than a dozen European banks switched from statistical-modeling approaches to Machine Learning tech- niques and, in some cases, increased their sales of new products by 10 percent and arXiv:1611.08903v1 [cs.LG] 27 Nov 2016 20 percent increases in cash collections [20]. Intelligent machines are used in a variety of domains, including writing news articles [5], finding promising recruits given their CV [9] and many more.
    [Show full text]
  • Applied Arts Department Cooks up a New Course
    2 THE NEW TRIER NEWS FEATURES FRIDAY, SEPTEMBER 13, 2013 Applied Arts Department cooks up a new course by Marissa Rogina didn’t have room for it.” Outside traditional recipes from these regions of culinary courses, Applied Arts and learn about how these cultures This year, New Trier’s Applied is also expanding other areas of and regions influence their cuisine. Arts Department added a new course its department. “You can see this Since this course is only offered to to its repertoire—International firsthand with our newer courses seniors it wraps up with a unit on Food. This class has been popular, such as Human Growth and Child cooking in college. featuring four class sections this Development, Project Lead the Cooking and baking are not the year. International Foods is different Way Engineering Courses, and our only things that students do to fill from the other culinary courses at Architectural course sequence,” their forty minutes in International New Trier—Creative Cuisine and continued Boumstein. Foods. They also spend time learning Gourmet Foods—because the foods Many students are confusing the about the various cultures and students cook are centered on what addition of International Foods with countries from which they cook up culture and influence is being studied Gourmet Foods, yet the two courses cuisine. Teachers initiate discussions in class. are actually very different. about these cultures and countries International Foods was Gourmet foods, offered only to that spark students to ask questions added because of the Applied Arts sophomores and juniors, focuses on and explore more about where the departments’ vision to offer more different types of gourmet cooking foods they make come from.
    [Show full text]
  • An Augmented Vision Tool with Real- Time Sensing Ability in Tagged Environments
    X-Vision: An Augmented Vision Tool with Real- Time Sensing Ability in Tagged Environments The MIT Faculty has made this article openly available. Please share how this access benefits you. Your story matters. Citation Sun, Yongbin et al. "X-Vision: An Augmented Vision Tool with Real-Time Sensing Ability in Tagged Environments." 2018 IEEE International Conference on RFID Technology & Application, September 2018, Macau, China, Institute of Electrical and Electronics Engineers (IEEE), December 2018. © 2018 IEEE As Published http://dx.doi.org/10.1109/rfid-ta.2018.8552778 Publisher Institute of Electrical and Electronics Engineers (IEEE) Version Author's final manuscript Citable link https://hdl.handle.net/1721.1/123808 Terms of Use Creative Commons Attribution-Noncommercial-Share Alike Detailed Terms http://creativecommons.org/licenses/by-nc-sa/4.0/ X-Vision: An augmented vision tool with real-time sensing ability in tagged environments Yongbin Sun* , Sai Nithin R. Kantareddy* , Rahul Bhattacharyya , and Sanjay E. Sarma Auto-ID Labs, Department of Mechanical Engineering Massachusetts Institute of Technology Cambridge, USA fyb sun, nithin, rahul b, sesarmag at mit.edu Abstract—We present the concept of X-Vision, an enhanced Augmented Reality (AR)-based visualization tool, with the real- time sensing capability in a tagged environment. We envision that this type of a tool will enhance the user-environment interaction and improve the productivity in factories, smart- spaces, home & office environments, maintenance/facility rooms and operation theatres, etc. In this paper, we describe the design of this visualization system built upon combining the object’s pose information estimated by the depth camera and the object’s ID & physical attributes captured by the RFID tags.
    [Show full text]
  • X-Vision: an Augmented Vision Tool with Real-Time Sensing Ability in Tagged Environments
    X-Vision: An augmented vision tool with real-time sensing ability in tagged environments Yongbin Sun* , Sai Nithin R. Kantareddy* , Rahul Bhattacharyya , and Sanjay E. Sarma Auto-ID Labs, Department of Mechanical Engineering Massachusetts Institute of Technology Cambridge, USA fyb sun, nithin, rahul b, sesarmag at mit.edu Abstract—We present the concept of X-Vision, an enhanced Augmented Reality (AR)-based visualization tool, with the real- time sensing capability in a tagged environment. We envision that this type of a tool will enhance the user-environment interaction and improve the productivity in factories, smart- spaces, home & office environments, maintenance/facility rooms and operation theatres, etc. In this paper, we describe the design of this visualization system built upon combining the object’s pose information estimated by the depth camera and the object’s ID & physical attributes captured by the RFID tags. We built a Fig. 1. Left: A user wearing the system sees a cup with overlaid temperature information. Right: System components: an Intel RealSense D415 RGB-D physical prototype of the system demonstrating the projection of camera is attached on a HoloLens via a custom mount. 3D holograms of the objects encoded with sensed information like water-level and temperature of common office/household objects. The paper also discusses the quality metrics used to compare the pose estimation algorithms for robust reconstruction of the attributes can be wirelessly retrieved and mapped to a digital object’s 3D data. avatar. In this paper, we have designed a visualization framework I. INTRODUCTION called X-Vision, hoping to equip users with the ability to di- Superimposing information on to the real-world, the concept rectly see the physical attributes of surrounding objects (Figure commonly known to us as Augmented reality (AR), has been 1).
    [Show full text]
  • 2015-Nmc-Technology-Outlook-Brazilian-Universities-PT.Pdf
    Panorama Tecnológico NMC 2015 Universidades Brasileiras Uma Análise Regional do Horizon Project Resumo Executivo ........................................................................................................................................................................................ 3 Introdução .............................................................................................................................................................................................................. 3 Principais Tendências que Aceleram a Adoção de Tecnologia ........................................................ 6 Desafios Significativos que Impedem a Adoção de Tecnologia ................................................. 10 Desenvolvimentos Importantes em Tecnologia Educacional Período para Adoção: 1 Ano ou Menos § Computação na Nuvem (Cloud Computing) ...................................................................................... 12 § Publicação Eletrônica ..................................................................................................................................... 13 § Aprendizagem On-line ................................................................................................................................. 14 § Redes Sociais ..................................................................................................................................................... 15 Período para Adoção: 2 a 3 Anos § Traga Seu Próprio Dispositivo (Bring Your Own Device) ..............................................................
    [Show full text]
  • L'avenir DE L'innovation Chroniques Digitales D'un Tour Du Monde
    L'AVENIR DE L'INNOVATION Chroniques digitales d'un tour du monde Francis Pisani Publié par l’Observatoire Netexplo © Netexplo. 2016 Tous droits réservés ISBN : 978-2-9546672-2-5 Note de l’auteur : Les textes mentionnés en bleu font référence à des informations complémentaires disponibles sur le web. La version numérique de ce livre, consultable sur www.netexplo.org, permet un accès direct à ces informations. A Colette, ma mère et à Henri, mon beau-père qui m’ont donné l’envie de connaître d’autres terres. A Edgard, mon père qui m’a aidé à comprendre qu’il venait d’ailleurs, comme nous tous. A Fabien, Émilia, et Yara mes enfants nomades. A Catherine, qui me donne le goût de voyages différents. A ceux, qui maintiennent la tête hors des « boîtes » dans lesquelles il est si facile de trouver ses aises… et à ceux, qui courent le risque de se mettre en marge, pratiquent la désobéissance créatrice, poussent le nez dehors et s’inspirent de ce qu’ils rencontrent pour mieux faire bouger le schmilblick… PARTIR … Mes tours du monde de l'innovation sont d'abord la continuation naturelle des 15 ans que j'ai passés au Mexique (de 1980 à 1995) et des 15 ans qui ont suivi en Californie (1996-2010). Dans un cas comme dans l'autre, j'étais en quête de ce qui peut changer le monde. J'avais choisi Mexico pour suivre les bouleversements d'Amérique centrale lancés par l'insurrection du Nicaragua (quand les sandinistes portaient encore un plus de démocratie). J'ai déménagé à San Francisco pour couvrir l’essor des technologies de l'information, telles qu'on les conçoit et qu'on les crée autour de Silicon Valley.
    [Show full text]