Neural Word Representations from Large-Scale Commonsense Knowledge

Total Page:16

File Type:pdf, Size:1020Kb

Neural Word Representations from Large-Scale Commonsense Knowledge Neural Word Representations from Large-Scale Commonsense Knowledge Jiaqiang Chen Niket Tandon Gerard de Melo IIIS, Tsinghua University Max Planck Institute for Informatics IIIS, Tsinghua University Beijing, China Saarbrucken,¨ Germany Beijing, China Email: [email protected] Email: [email protected] Email: [email protected] Greek are likely related. Our training objective thus encourages saliently related words to have similar representations. Abstract—There has recently been a surge of research on neural network-inspired algorithms to produce numerical vector representations of words, based on contextual information. In this II. RELATED WORK paper, we present an approach to improve such word embeddings Many of the current methods for obtaining distributed word by first mining cognitively salient word relationships from text and then using stochastic gradient descent to jointly optimize embeddings are neural network-inspired and aim at rather the embeddings to reflect this information, in addition to the dense real-valued vector spaces. While early work focused regular contextual information captured by the word2vec CBOW on probabilistic language models [3], Collobert et al. [4] used objective. Our findings show that this new training regime leads to a convolutional neural network to maximize the difference vectors that better reflect commonsense information about words. between scores from text windows in a large training corpus and corresponding randomly generated negative examples. Mikolov et al. [1] proposed simplified network architectures I. INTRODUCTION to efficiently train such vectors at a much faster rate and thus also at a much larger scale. Their word2vec1 implementation Words are substantially discrete in nature, and hence, tradi- provides two architectures, the CBOW and the Skip-gram tionally, the vast majority of natural language processing tools, models. CBOW also relies on a window approach, attempting including statistical ones, have regarded words as distinct atomic to use the surrounding words to predict the current target word. symbols. In recent years, however, the idea of embedding words However, it simplifies the hidden layer to be just the average in a vector space using neural network-inspired algorithms has of surrounding words’ embeddings. The Skip-gram model tries gained enormous popularity. Mapping words to vectors in a to do the opposite. It uses the current word to predict the way that reflects word similarities provides machine learning surrounding words. In our approach, we build on the CBOW algorithms with much-needed generalization ability. If the words variant, as its optimization runs faster. car and automobile have similar vectors, a learning algorithm is better equipped to generalize from one word to the other. There have been other proposals to adapt the word2vec Word embeddings are typically trained using large amounts of models. Levy et al. [5] use dependency parse relations to contextual data. While regular sentence contexts play a vital create word embeddings that are able to capture contextual role in meaning acquisition, words and concepts are often also relationships between words that are further apart in the acquired by other means. Humans may pay special attention sentence. Further analysis revealed that their word embeddings to certain cognitively salient features of an object, or rely on capture more functional but less topical similarity. Faruqui et more explicit definitions (e.g., looking up a meaning online). al. [6] apply post-processing steps to existing word embeddings in order to bring them more in accordance with semantic In this paper, we propose a model to jointly train word lexicons. Rather than using rich structured knowledge sources, representations not just on regular contexts, as in the word2vec our work focuses on improving word embeddings using textual CBOW model [1], but also to reflect more salient information. data, by relying on information extraction to expose particularly For the latter, we use information extraction techniques [2] valuable contexts and relationships in a text corpus. In particular, on large-scale text data to mine definitions and synonyms we are not aware of any previous work that mines large-scale as well as lists and enumerations. Rather than considering all common-sense knowledge to train embeddings of lexical units. contexts as equal, our approach can be viewed as treating certain specific contexts as more informative than others. Consider III. SALIENT PROXIMITY MODEL the sentence The Roman Empire was remarkably multicultural, with ”a rather astonishing cohesive capacity” to create a sense Our approach is to simultaneously train the word embed- of shared identity.... While it contains several useful signals, dings on generic contexts from the corpus on the one hand and Roman does not seem to bear an overly close relationship with on semantically significant contexts, obtained using extraction capacity, astonishing, or sense. In contrast, upon encountering techniques, on the other hand. For the regular general contexts, Greek and Roman mythology, we may conclude that Roman and we draw on the word2vec CBOW model [1] to predict a word given its surrounding neighbors in the corpus. This research was supported by China 973 Program Grants 2011CBA00300, 2011CBA00301, and NSFC Grants 61033001, 61361136003, 61450110088. 1https://code.google.com/p/word2vec/ At the same time, our model relies on our ability to extract as negative sample pairs. We attempt to maximize the score semantically salient contexts that are more indicative of word for the positive training data and minimize the score of the meanings. These extractions will be described in detail later negative samples. In the training procedure, this amounts to in Section IV. Our algorithm assumes that they have been simply generating k random negative samples for each extracted transformed into a set of word pairs likely to be closely related, word pair. That is, we replace wr with random words from which are used to modify the word embeddings. Due to this the vocabulary. For the negative samples, we assign the label more focused information, we expect the final word embeddings l = 0, whereas for the original word pairs, l = 1. Now, for to reflect more semantic information than embeddings trained each word pair we try to minimize its loss L: only on regular contexts. Given an extracted pair of related words, the intuition is that the embeddings for the two words L = −l log f − (1 − l) log(1 − f) (5) f = σ(vT · v ) should be pulled together. Given a word wt, our objective wt wr (6) function attempts to maximize the probability of finding its related words wr: We use stochastic gradient descent to optimize this function. T The formulae for the gradient are easy to compute: 1 X X log p(w jw ) (1) T r t @L 1 1 t=1 wr = −l f (1 − f) Vwt + (1 − l) f (1 − f) Vwt @Vwr f 1 − f Here, T is the vocabulary size and the probabilities are modeled = −(l − f) Vw (7) using a softmax, defined as follows: t @L T = −(l − f) V (8) exp(V · V ) wr wr wr @V p(w jw ) = (2) wt r t P exp(V T · V ) w 0 wr0 wt r This objective is optimized alongside with the original V , V refer to the word vectors for two related words w , wr wt r word2vec CBOW objective, i.e., our overall model combines w , while w 0 with corresponding vectors V refer to all t r wr0 the two objectives. Implementation-wise, we train the model possible words. We use the inner product to score how well in parallel with the CBOW model, which allows us to inject two words match. When they are very similar or related, their the extracted knowledge into the word vectors such that they embeddings should be close to each other and hence the inner are reflected during the CBOW training rather than just as a product of their embeddings should be large. Using the softmax post-processing step. Thus we obtain a joint learning process function, we can take a maximum likelihood approach to train in which the two components are able to mutually influence the embeddings in a tractable manner. However, the gradient each other. Both objectives contribute to the embeddings’ is as follows: ability to capture semantic relationships. Training with the T P T @V ·V log w exp(Vw ·Vw ) @ log p(wrjwt) wr wt r0 r0 t extracted contexts enables us to adjust word embeddings based @V = @V − @V wt wt wt on concrete evidence of semantic relationships, while the use of exp(Vw ·Vw ) P r0 r general corpus contexts enables us to maintain the advantages = Vwr − 0 P T Vwr0 wr w exp(Vw ·Vwt ) P r0 r0 of the word2vec CBOW model, in particular its ability to benefit = Vw − p(wr0 jwt)Vw 0 r wr0 r from massive volumes of raw corpus data. = Vwr − EpVwr0 (3) IV. INFORMATION EXTRACTION To compute this gradient, the expectation of all the word vectors with respect to their probabilities would be needed. The Our model can flexibly incorporate semantic relationships time complexity for this is proportional to the vocabulary size, extracted using various kinds of information extraction methods. which is typically very large. Here, we use negative sampling Different kinds of sources and extraction methods can bring as a speed-up technique [7]. This can be viewed as a simplified different sorts of information to the vectors, suitable for different version of Noise Contrastive Estimation (NCE) [8], which applications. In our experiments, we investigate two main reduces the problem of determining the softmax to that of sources: a dictionary corpus from which we extract definitions binary classification, discriminating between samples from the and synonyms, and a general Web corpus, from which we data distribution and negative samples. In particular, we consider extract lists. Our model could similarly be used with other a distribution of random noise and optimize for discriminating extraction methods.
Recommended publications
  • Multilingual Ontology Acquisition from Multiple Mrds
    Multilingual Ontology Acquisition from Multiple MRDs Eric Nichols♭, Francis Bond♮, Takaaki Tanaka♮, Sanae Fujita♮, Dan Flickinger ♯ ♭ Nara Inst. of Science and Technology ♮ NTT Communication Science Labs ♯ Stanford University Grad. School of Information Science Natural Language ResearchGroup CSLI Nara, Japan Keihanna, Japan Stanford, CA [email protected] {bond,takaaki,sanae}@cslab.kecl.ntt.co.jp [email protected] Abstract words of a language, let alone those words occur- ring in useful patterns (Amano and Kondo, 1999). In this paper, we outline the develop- Therefore it makes sense to also extract data from ment of a system that automatically con- machine readable dictionaries (MRDs). structs ontologies by extracting knowledge There is a great deal of work on the creation from dictionary definition sentences us- of ontologies from machine readable dictionaries ing Robust Minimal Recursion Semantics (a good summary is (Wilkes et al., 1996)), mainly (RMRS). Combining deep and shallow for English. Recently, there has also been inter- parsing resource through the common for- est in Japanese (Tokunaga et al., 2001; Nichols malism of RMRS allows us to extract on- et al., 2005). Most approaches use either a special- tological relations in greater quantity and ized parser or a set of regular expressions tuned quality than possible with any of the meth- to a particular dictionary, often with hundreds of ods independently. Using this method, rules. Agirre et al. (2000) extracted taxonomic we construct ontologies from two differ- relations from a Basque dictionary with high ac- ent Japanese lexicons and one English lex- curacy using Constraint Grammar together with icon.
    [Show full text]
  • Latexsample-Thesis
    INTEGRAL ESTIMATION IN QUANTUM PHYSICS by Jane Doe A dissertation submitted to the faculty of The University of Utah in partial fulfillment of the requirements for the degree of Doctor of Philosophy Department of Mathematics The University of Utah May 2016 Copyright c Jane Doe 2016 All Rights Reserved The University of Utah Graduate School STATEMENT OF DISSERTATION APPROVAL The dissertation of Jane Doe has been approved by the following supervisory committee members: Cornelius L´anczos , Chair(s) 17 Feb 2016 Date Approved Hans Bethe , Member 17 Feb 2016 Date Approved Niels Bohr , Member 17 Feb 2016 Date Approved Max Born , Member 17 Feb 2016 Date Approved Paul A. M. Dirac , Member 17 Feb 2016 Date Approved by Petrus Marcus Aurelius Featherstone-Hough , Chair/Dean of the Department/College/School of Mathematics and by Alice B. Toklas , Dean of The Graduate School. ABSTRACT Blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah. Blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah. Blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah. Blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah. Blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah. Blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah. Blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah. Blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah.
    [Show full text]
  • Linux-Cookbook.Pdf
    LINUX COOKBOOK ™ Other Linux resources from O’Reilly Related titles Linux Device Drivers Exploring the JDS Linux Linux in a Nutshell Desktop Running Linux Learning Red Hat Enterprise Building Embedded Linux Linux and Fedora Systems Linux Pocket Guide Linux Security Cookbook Understanding the Linux Kernel Linux Books linux.oreilly.com is a complete catalog of O’Reilly’s books on Resource Center Linux and Unix and related technologies, including sample chapters and code examples. ONLamp.com is the premier site for the open source web plat- form: Linux, Apache, MySQL, and either Perl, Python, or PHP. Conferences O’Reilly brings diverse innovators together to nurture the ideas that spark revolutionary industries. We specialize in document- ing the latest tools and systems, translating the innovator’s knowledge into useful skills for those in the trenches. Visit conferences.oreilly.com for our upcoming events. Safari Bookshelf (safari.oreilly.com) is the premier online refer- ence library for programmers and IT professionals. Conduct searches across more than 1,000 books. Subscribers can zero in on answers to time-critical questions in a matter of seconds. Read the books on your Bookshelf from cover to cover or sim- ply flip to the page you need. Try it today with a free trial. LINUX COOKBOOK ™ Carla Schroder Beijing • Cambridge • Farnham • Köln • Paris • Sebastopol • Taipei • Tokyo Linux Cookbook™ by Carla Schroder Copyright © 2005 O’Reilly Media, Inc. All rights reserved. Printed in the United States of America. Published by O’Reilly Media, Inc., 1005 Gravenstein Highway North, Sebastopol, CA 95472. O’Reilly books may be purchased for educational, business, or sales promotional use.
    [Show full text]
  • Pipenightdreams Osgcal-Doc Mumudvb Mpg123-Alsa Tbb
    pipenightdreams osgcal-doc mumudvb mpg123-alsa tbb-examples libgammu4-dbg gcc-4.1-doc snort-rules-default davical cutmp3 libevolution5.0-cil aspell-am python-gobject-doc openoffice.org-l10n-mn libc6-xen xserver-xorg trophy-data t38modem pioneers-console libnb-platform10-java libgtkglext1-ruby libboost-wave1.39-dev drgenius bfbtester libchromexvmcpro1 isdnutils-xtools ubuntuone-client openoffice.org2-math openoffice.org-l10n-lt lsb-cxx-ia32 kdeartwork-emoticons-kde4 wmpuzzle trafshow python-plplot lx-gdb link-monitor-applet libscm-dev liblog-agent-logger-perl libccrtp-doc libclass-throwable-perl kde-i18n-csb jack-jconv hamradio-menus coinor-libvol-doc msx-emulator bitbake nabi language-pack-gnome-zh libpaperg popularity-contest xracer-tools xfont-nexus opendrim-lmp-baseserver libvorbisfile-ruby liblinebreak-doc libgfcui-2.0-0c2a-dbg libblacs-mpi-dev dict-freedict-spa-eng blender-ogrexml aspell-da x11-apps openoffice.org-l10n-lv openoffice.org-l10n-nl pnmtopng libodbcinstq1 libhsqldb-java-doc libmono-addins-gui0.2-cil sg3-utils linux-backports-modules-alsa-2.6.31-19-generic yorick-yeti-gsl python-pymssql plasma-widget-cpuload mcpp gpsim-lcd cl-csv libhtml-clean-perl asterisk-dbg apt-dater-dbg libgnome-mag1-dev language-pack-gnome-yo python-crypto svn-autoreleasedeb sugar-terminal-activity mii-diag maria-doc libplexus-component-api-java-doc libhugs-hgl-bundled libchipcard-libgwenhywfar47-plugins libghc6-random-dev freefem3d ezmlm cakephp-scripts aspell-ar ara-byte not+sparc openoffice.org-l10n-nn linux-backports-modules-karmic-generic-pae
    [Show full text]
  • Integral Estimation in Quantum Physics
    INTEGRAL ESTIMATION IN QUANTUM PHYSICS by Jane Doe A dissertation submitted to the faculty of The University of Utah in partial fulfillment of the requirements for the degree of Doctor of Philosophy in Mathematical Physics Department of Mathematics The University of Utah May 2016 Copyright c Jane Doe 2016 All Rights Reserved The University of Utah Graduate School STATEMENT OF DISSERTATION APPROVAL The dissertation of Jane Doe has been approved by the following supervisory committee members: Cornelius L´anczos , Chair(s) 17 Feb 2016 Date Approved Hans Bethe , Member 17 Feb 2016 Date Approved Niels Bohr , Member 17 Feb 2016 Date Approved Max Born , Member 17 Feb 2016 Date Approved Paul A. M. Dirac , Member 17 Feb 2016 Date Approved by Petrus Marcus Aurelius Featherstone-Hough , Chair/Dean of the Department/College/School of Mathematics and by Alice B. Toklas , Dean of The Graduate School. ABSTRACT Blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah. Blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah. Blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah. Blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah. Blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah. Blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah. Blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah. Blah blah blah blah blah blah blah blah blah blah blah blah blah blah blah.
    [Show full text]
  • 12. Practical Programming for NLP Patrick Jeuniaux1, Andrew Olney2, Sidney D'mello2 1 Université Laval, Canada, 2 University of Memphis, USA
    12. Practical Programming for NLP Patrick Jeuniaux1, Andrew Olney2, Sidney D'Mello2 1 Université Laval, Canada, 2 University of Memphis, USA ABSTRACT This chapter is aimed at students and researchers who are eager to learn about practical programmatic solutions to natural language processing (NLP) problems. In addition to introducing the readers to programming basics, programming tools, and complete programs, we also hope to pique their interest to actively explore the broad and fascinating field of automatic natural language processing. Part I introduces programming basics and the Python programming language. Part II takes a step by step approach in illustrating the development of a program to solve a NLP problem. Part III provides some hints to help readers initiate their own NLP programming projects. INTRODUCTION Natural language processing (NLP) attempts to automatically analyze the languages spoken by humans (i.e., natural languages). For instance, you can program a computer to automatically identify the language of a text, extract the grammatical structure of sentences (see Chapter XXX of this book), categorize texts by genre (e.g., decide whether a text is a scientific or a narrative text; see Chapter XXX for classification applications), summarize a book (see Chapter XXX), etc. This chapter is aimed at teaching specialized, yet introductory, programming skills that are required to use available NLP tools. We hope that this chapter serves as a catalyst to launch NLP projects by motivating novice programmers to learn more about programming and encouraging more advanced programmers to develop NLP programs. The chapter is aimed at readers from the interdisciplinary arena that encompasses computer science, cognitive psychology, and linguistics.
    [Show full text]
  • A Visual Approach for Text Analysis Using Multiword Topics Seongmin Mun, Guillaume Desagulier, Kyungwon Lee
    A visual approach for text analysis using multiword topics Seongmin Mun, Guillaume Desagulier, Kyungwon Lee To cite this version: Seongmin Mun, Guillaume Desagulier, Kyungwon Lee. A visual approach for text analysis using multiword topics. EuroVis 2017, Eurographics, Jun 2017, Barcelona, Spain. 10.2312/eurp.20171168. halshs-01590990 HAL Id: halshs-01590990 https://halshs.archives-ouvertes.fr/halshs-01590990 Submitted on 20 Sep 2017 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Eurographics Conference on Visualization (EuroVis), Posters Track (2017) A. Puig Puig and T. Isenberg (Editors) A visual approach for text analysis using multiword topics Seongmin Mun1;2 , Guillaume Desagulier3, Kyungwon Lee4 1Lifemedia Interdisciplinary Program, Ajou University, South Korea 2UMR 7114 MoDyCo - CNRS, University Paris Nanterre, France 3UMR 7114 MoDyCo - University Paris 8, CNRS, University Paris Nanterre 4Department of Digital Media, Ajou University, South Korea Abstract Topics in a text corpus include features and information; visualizing these topics can improve a user’s understanding of the corpus. Topics can be broadly divided into two categories: those whose meaning can be described in one word and those whose meaning in expressed through a combination of words.
    [Show full text]
  • Mac OS X DICT Howto BETA
    Mac OS X DICT Howto BETA How to get a superior & free dictionary up and running on your Mac a step-by-step walkthrough by Jens Ropers (www.ropersonline.com) -> Please jump to Part II if you don’t care for chatty introductions. -> Please note that this document is in BETA stage. See the appendix for more info. Part I: Introduction Let me own up: I, too, am a switcher. And back and forth, too - but here I am again, writing on my iBook. ;-) Aaanyways, on Windows I got to know a nice program named WordWeb. A free dictionary. And none too shabby! (http://wordweb.info/) I couldn’t find one like that for the Mac. I emailed the author, would he consider porting to Mac ? He wouldn’t. (The good man had probably never touched a Mac, and was needlessly scared.) I then did some research on my own and found out that WordWeb was actually using a database from Princeton University called WordNet. (http://www.cogsci.princeton.edu/~wn/) Quite some time later, I came across the OmniDictionary program from the Omni Group. (http://www.omnigroup.com/) It could access public dictionary servers from the DICT Development Group. (http://www.dict.org/) To my delight I discovered that the WordNet database (among others) was being used there as well. To my dismay I discovered that this only worked as long as I was connected to the Internet. What to do? It was mentioned in various places that it should be possible to get programs like OmniDictionary to work offline, i.e.
    [Show full text]
  • El Sistema Operativo Linux
    El Sistema Operativo Linux Javier Parapar Contenido El Sistema Operativo Linux Conceptos B´asicos Javier Parapar ([email protected]) Aula de Formaci´onInform´atica Universidade da Coru˜na Material base por Carlos Abalde 8 - 17 de enero de 2007 Javier Parapar (AFI) El Sistema Operativo Linux 8 - 17 de enero de 2007 1 / 159 Contenido El Sistema Operativo Linux Javier Parapar 1 El software libre y Linux. Distribuciones Contenido 2 Primeros pasos en Linux 3 Instalaci´onde distribuciones 4 Gesti´onde archivos (I) 5 Gesti´onde archivos (y II) 6 Edici´onde archivos de texto 7 Gesti´onde usuarios y procesos 8 Shell scripts 9 Arranque, reinicio y apagado del sistema 10 Logs del sistema 11 Sistema gr´aficoXwindow 12 Configuraci´onb´asicadel entorno 13 Sistemas de paquetes. Compilaci´onde software no empaquetado 14 Acceso y servicios de Internet 15 Paquetes ofim´aticos.Multimedia 16 Otros conceptos avanzados Javier Parapar (AFI) El Sistema Operativo Linux 8 - 17 de enero de 2007 2 / 159 Parte I El Sistema Operativo Linux Javier Parapar El software libre y Linux El software libre y Linux. Distribuciones Los inicios La licencia GPL GNU/Linux Linux hoy 1 El software libre y Linux Distribuciones Los inicios Slackware Debian GNU/Linux La licencia GPL Knoppix Ubuntu GNU/Linux Fedora Core Linux hoy SuSe Mandriva Gentoo Linux 2 Distribuciones Slackware Debian GNU/Linux Knoppix Ubuntu Fedora Core SuSe Mandriva Gentoo Linux Javier Parapar (AFI) El Sistema Operativo Linux 8 - 17 de enero de 2007 3 / 159 Parte I El Sistema Operativo Linux Javier Parapar El software libre y Linux El software libre y Linux.
    [Show full text]
  • Unix Et Programmation Shell
    Unix et Programmation Shell Philippe Langevin d´epartment d'informatique UFR sciences et technique universit´edu sud Toulon Var Automne 2013 Philippe Langevin (imath/ustv) Unix et Programmation Shell Automne 2013 1 / 33 document brouillon en r´evision site du cours : http://langevin.univ-tln.fr/cours/UPS/upsh.html localisation du fichier : http://langevin.univ-tln.fr/cours/UPS/doc/intro.pdf Philippe Langevin (imath/ustv) Unix et Programmation Shell Automne 2013 2 / 33 derni`eresmodifications intro.tex 2013-06-23 10:07:23.697660496 +0200 macros.tex 2013-06-22 23:42:16.868263946 +0200 prologue.tex 2013-06-22 22:10:45.967471216 +0200 upsh.tex 2013-06-22 11:24:27.941744751 +0200 tools.tex 2013-06-22 11:24:27.940741957 +0200 term.tex 2013-06-22 11:24:27.934741399 +0200 syntaxe.tex 2013-06-22 11:24:27.931741678 +0200 proc.tex 2013-06-22 11:24:27.927741678 +0200 piped.tex 2013-06-22 11:24:27.922741399 +0200 perm.tex 2013-06-22 11:24:27.921741678 +0200 part.tex 2013-06-22 11:24:27.918742237 +0200 man.tex 2013-06-22 11:24:27.911741957 +0200 langage.tex 2013-06-23 10:04:58.364659987 +0200 file.tex 2013-06-22 11:24:27.896741119 +0200 dup.tex 2013-06-22 11:24:27.885741678 +0200 fic.tex 2013-06-22 11:24:27.885741678 +0200 bash.tex 2013-06-22 11:24:27.881741957 +0200 Philippe Langevin (imath/ustv) Unix et Programmation Shell Automne 2013 3 / 33 shell unix 1 - shell unix origine unices GNU/linux distribution shell unix GUI vs CLI exemple Philippe Langevin (imath/ustv) Unix et Programmation Shell Automne 2013 4 / 33 1969 Ken Thompson cr´eele syst`eme UNICS 1971 Dennis Ritchie cr´ee Le langage C utile pour le d´eveloppement d'unix.
    [Show full text]
  • GNU Dico Manual Version 2.10, 4 September 2020
    GNU Dico Manual version 2.10, 4 September 2020 Sergey Poznyakoff. Published by the Free Software Foundation, 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA Copyright c 2008–2020 Sergey Poznyakoff Permission is granted to copy, distribute and/or modify this document un- der the terms of the GNU Free Documentation License, Version 1.3 or any later version published by the Free Software Foundation; with no Invariant Sections, no Front-Cover, and no Back-Cover texts. A copy of the license is included in the section entitled “GNU Free Documentation License”. D´edi´e`ala m´emoire de Jacques Brel. i Short Contents Preface ................................................. 1 1 Overview ............................................ 3 2 Introduction to GNU Dico .............................. 5 3 Building the Package .................................. 7 4 The dicod daemon. ................................... 9 5 Modules ............................................ 43 6 Dico Module Interface ................................ 71 7 Dico — a client program. .............................. 81 8 GCIDER ........................................... 99 9 How to Report a Bug ................................ 101 A Available Strategies ................................. 103 B Dictionary Server Protocol ............................ 105 C Time and Date Formats .............................. 113 D The Libdico Library ................................. 117 E GNU Free Documentation License ...................... 133 Concept Index ........................................
    [Show full text]
  • Open Translation Tools: Disruptive Potential to Broaden Access to Knowledge
    Open Translation Tools: Disruptive Potential to Broaden Access to Knowledge Prepared for Open Society Institute by Allen Gunn Aspiration September 2008 This document is distributed under a Creative Commons Attribution-ShareAlike 3.0 License San Francisco Nonprofit Technology Center, 1370 Mission St., San Francisco, CA 94103 Phone: (415) 839-6456 • [email protected] • Page 1 Table of Contents 1. Executive Summary..........................................................................3 2. Acknowledgments............................................................................6 3. OTT07 Event Background and Agenda...........................................7 4. Translation Concepts......................................................................11 5. Use Cases for Open Content Translation......................................13 6. Open Translation Tools..................................................................17 7. OTT07 “Dream” Translation Tool...................................................22 8. Technical Innovations and Considerations....................................26 9. Related Issues and Gating Factors................................................33 10.Event Outcomes and Ideas for Next Steps....................................36 11.Appendix I: Open Translation Use Cases......................................39 12.Appendix II: Open Translation Tools..............................................49 San Francisco Nonprofit Technology Center, 1370 Mission St., San Francisco, CA 94103 Phone: (415) 839-6456 •
    [Show full text]