Neuron C Reference Guide Iii • Introduction to the LONWORKS Platform (078-0391-01A)

Total Page:16

File Type:pdf, Size:1020Kb

Neuron C Reference Guide Iii • Introduction to the LONWORKS Platform (078-0391-01A) Neuron C Provides reference info for writing programs using the Reference Guide Neuron C programming language. 078-0140-01G Echelon, LONWORKS, LONMARK, NodeBuilder, LonTalk, Neuron, 3120, 3150, ShortStack, LonMaker, and the Echelon logo are trademarks of Echelon Corporation that may be registered in the United States and other countries. Other brand and product names are trademarks or registered trademarks of their respective holders. Neuron Chips and other OEM Products were not designed for use in equipment or systems, which involve danger to human health or safety, or a risk of property damage and Echelon assumes no responsibility or liability for use of the Neuron Chips in such applications. Parts manufactured by vendors other than Echelon and referenced in this document have been described for illustrative purposes only, and may not have been tested by Echelon. It is the responsibility of the customer to determine the suitability of these parts for each application. ECHELON MAKES AND YOU RECEIVE NO WARRANTIES OR CONDITIONS, EXPRESS, IMPLIED, STATUTORY OR IN ANY COMMUNICATION WITH YOU, AND ECHELON SPECIFICALLY DISCLAIMS ANY IMPLIED WARRANTY OF MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. No part of this publication may be reproduced, stored in a retrieval system, or transmitted, in any form or by any means, electronic, mechanical, photocopying, recording, or otherwise, without the prior written permission of Echelon Corporation. Printed in the United States of America. Copyright © 2006, 2014 Echelon Corporation. Echelon Corporation www.echelon.com Welcome This manual describes the Neuron® C Version 2.3 programming language. It is a companion piece to the Neuron C Programmer's Guide. It provides reference information for writing programs using Neuron C. Neuron C is a programming language based on ANSI C that is designed for applications that run on Neuron Chips and Smart Transceivers (Neuron-hosted devices). Neuron C includes network communication, I/O, and event-handling extensions to ANSI C, which make it a powerful tool for the development of LONWORKS® applications. A subset of the Neuron C language is also used to describe the interoperable interface of host-based applications that are designed with the ShortStack® Developer’s Kit, FTXL™ Developer’s Kit, or the i.LON® SmartServer. This interoperable interface is contained within a file called a model file, which contains Neuron C declarations and definitions for the device interface. This guide focuses on the Neuron C language used for Neuron-hosted application development, and only highlights differences for model file compilation where necessary. Audience The Neuron C Programmer’s Guide is intended for application programmers who are developing LONWORKS applications. Readers of this guide are assumed to be familiar with the ANSI C programming language, and have some C programming experience. For a complete description of ANSI C, consult the following references: • —. 1989. American National Standard for Information Systems Programming Language C. Standard number X3.159-1989. New York, NY: American National Standards Institute. • —. 2007. International Standard ISO/IEC 9899:1999. Programming languages – C. Geneva, Switzerland: International Organization for Standardization. • Harbison, Samuel P. and Guy L. Steele, Jr. 2002. C: A Reference Manual, 5th edition. Upper Saddle River, NJ: Prentice Hall, Inc. • Kernighan, Brian W. and Dennis M. Ritchie. 1988. The C Programming Language, 2nd edition. Upper Saddle River, NJ: Prentice Hall, Inc. • Plauger, P.J. and Jim Brodie. 1989. Standard C: Programmer’s Quick Reference Series. Buffalo, NY: Microsoft Press. • Plauger, P.J. and Jim Brodie. 1992. ANSI and ISO Standard C Programmer's Reference. Buffalo, NY: Microsoft Press. Related Documentation The following manuals are available from the Echelon Web site (www.echelon.com) and provide additional information that can help you develop Neuron C applications for LONWORKS devices: Neuron C Reference Guide iii • Introduction to the LONWORKS Platform (078-0391-01A). This manual provides an introduction to the ISO/IEC 14908 (ANSI/CEA-709.1 and EN14908) Control Network Protocol, and provides a high-level introduction to LONWORKS networks and the tools and components that are used for developing, installing, operating, and maintaining them. • I/O Model Reference for Smart Transceivers and Neuron Chips (078- 0392-01C). This manual describes the I/O models that are available for Echelon’s Smart Transceivers and Neuron Chips. • IzoT Commissioning Tool User's Guide (078-0514-01 ). This manual describes how to use the Izot Commissioning Tool to design, commission, monitor and control, maintain, and manage a LONWORKS network. • LONMARK® Application Layer Interoperability Guidelines. This manual describes design guidelines for developing applications for open interoperable LONWORKS devices, and is available from the LONMARK Web site, www.lonmark.org. • Neuron C Programmer’s Guide (078-0002-01I). This manual describes how to write programs using the Neuron C Version 2.3 programming language. • Neuron Tools Errors Guide (078-0402-01D). This manual documents and explains the various warning and error messages that can occur for the various Neuron C development tools. • IzoT NodeBuilder FX User’s Guide ( 078-0516-01). This manual describes how to develop a LONWORKS device using the IzoT NodeBuilder tool. All of the Echelon documentation is available in Adobe PDF format. To view the PDF files, you must have a current version of the Adobe Reader, which you can download from Adobe at: www.adobe.com/products/acrobat/readstep2.html. Typographic Conventions for Syntax Table 1 lists the typographic conventions used in this manual for displaying Neuron C syntax: Table 1. Typographic Conventions Typeface or Symbol Used for Example boldface type keywords network literal characters { italic type abstract elements identifier [square brackets] optional fields [bind-info] iv Typeface or Symbol Used for Example | vertical bar a choice between two input | output elements Example: The syntax for declaring a network variable is: network input | output [netvar modifier] [class] type [bind-info] identifier • You type the keywords network, input, and output as shown • You replace the abstract elements netvar modifier, class, type, bind-info, and identifier with the actual modifier, class, type, bind information, and identifier for the network variable • The declaration must include either input or output, but not both • The elements netvar modifier, class, and bind-info are all optional When a particular element or expression includes punctuation, such as quotation marks, parentheses, and semicolons (but not including square brackets and vertical bars), you must type that punctuation as shown. Code examples appear in the monospace Courier font: #include <mem.h> unsigned array1[40], array2[40]; // See if array1 matches array2 if (memcmp(array1, array2, 40) != 0) { // The contents of the two areas do not match } Neuron C Reference Guide v Table of Contents Welcome ......................................................................................................... iii Audience ........................................................................................................ iii Related Documentation ................................................................................ iii Typographic Conventions for Syntax ........................................................... iv Neuron C Overview ..................................................................................... xix Chapter 1. Predefined Events ........................................................................ 1 Introduction to Predefined Events ................................................................ 2 Event Directory .............................................................................................. 3 flush_completes Event............................................................................. 3 Syntax ................................................................................................ 3 Example ............................................................................................. 3 io_changes Event ..................................................................................... 3 Syntax ................................................................................................ 4 Example 1 .......................................................................................... 4 Example 2 .......................................................................................... 5 io_in_ready Event .................................................................................... 5 Syntax ................................................................................................ 5 Example ............................................................................................. 5 io_out_ready Event .................................................................................. 5 Syntax ................................................................................................ 5 Example ............................................................................................. 5 io_update_occurs Event ........................................................................... 6 Syntax ................................................................................................ 6 Example ............................................................................................
Recommended publications
  • Spiking Neural Networks: Neuron Models, Plasticity, and Graph Applications
    Virginia Commonwealth University VCU Scholars Compass Theses and Dissertations Graduate School 2015 Spiking Neural Networks: Neuron Models, Plasticity, and Graph Applications Shaun Donachy Virginia Commonwealth University Follow this and additional works at: https://scholarscompass.vcu.edu/etd Part of the Theory and Algorithms Commons © The Author Downloaded from https://scholarscompass.vcu.edu/etd/3984 This Thesis is brought to you for free and open access by the Graduate School at VCU Scholars Compass. It has been accepted for inclusion in Theses and Dissertations by an authorized administrator of VCU Scholars Compass. For more information, please contact [email protected]. c Shaun Donachy, July 2015 All Rights Reserved. SPIKING NEURAL NETWORKS: NEURON MODELS, PLASTICITY, AND GRAPH APPLICATIONS A Thesis submitted in partial fulfillment of the requirements for the degree of Master of Science at Virginia Commonwealth University. by SHAUN DONACHY B.S. Computer Science, Virginia Commonwealth University, May 2013 Director: Krzysztof J. Cios, Professor and Chair, Department of Computer Science Virginia Commonwealth University Richmond, Virginia July, 2015 TABLE OF CONTENTS Chapter Page Table of Contents :::::::::::::::::::::::::::::::: i List of Figures :::::::::::::::::::::::::::::::::: ii Abstract ::::::::::::::::::::::::::::::::::::: v 1 Introduction ::::::::::::::::::::::::::::::::: 1 2 Models of a Single Neuron ::::::::::::::::::::::::: 3 2.1 McCulloch-Pitts Model . 3 2.2 Hodgkin-Huxley Model . 4 2.3 Integrate and Fire Model . 6 2.4 Izhikevich Model . 8 3 Neural Coding Techniques :::::::::::::::::::::::::: 14 3.1 Input Encoding . 14 3.1.1 Grandmother Cell and Distributed Representations . 14 3.1.2 Rate Coding . 15 3.1.3 Sine Wave Encoding . 15 3.1.4 Spike Density Encoding . 15 3.1.5 Temporal Encoding . 16 3.1.6 Synaptic Propagation Delay Encoding .
    [Show full text]
  • Navigation Techniques in Augmented and Mixed Reality: Crossing the Virtuality Continuum
    Chapter 20 NAVIGATION TECHNIQUES IN AUGMENTED AND MIXED REALITY: CROSSING THE VIRTUALITY CONTINUUM Raphael Grasset 1,2, Alessandro Mulloni 2, Mark Billinghurst 1 and Dieter Schmalstieg 2 1 HIT Lab NZ University of Canterbury, New Zealand 2 Institute for Computer Graphics and Vision Graz University of Technology, Austria 1. Introduction Exploring and surveying the world has been an important goal of humankind for thousands of years. Entering the 21st century, the Earth has almost been fully digitally mapped. Widespread deployment of GIS (Geographic Information Systems) technology and a tremendous increase of both satellite and street-level mapping over the last decade enables the public to view large portions of the 1 2 world using computer applications such as Bing Maps or Google Earth . Mobile context-aware applications further enhance the exploration of spatial information, as users have now access to it while on the move. These applications can present a view of the spatial information that is personalised to the user’s current context (context-aware), such as their physical location and personal interests. For example, a person visiting an unknown city can open a map application on her smartphone to instantly obtain a view of the surrounding points of interest. Augmented Reality (AR) is one increasingly popular technology that supports the exploration of spatial information. AR merges virtual and real spaces and offers new tools for exploring and navigating through space [1]. AR navigation aims to enhance navigation in the real world or to provide techniques for viewpoint control for other tasks within an AR system. AR navigation can be naively thought to have a high degree of similarity with real world navigation.
    [Show full text]
  • Scripting NEURON
    Scripting NEURON Robert A. McDougal Yale School of Medicine 7 August 2018 What is a script? A script is a file with computer-readable instructions for performing a task. In NEURON, scripts can: set-up a model, define and perform an experimental protocol, record data, . Why write scripts for NEURON? Automation ensures consistency and reduces manual effort. Facilitates comparing the suitability of different models. Facilitates repeated experiments on the same model with different parameters (e.g. drug dosages). Facilitates recollecting data after change in experimental protocol. Provides a complete, reproducible version of the experimental protocol. Programmer's Reference neuron.yale.edu Use the \Switch to HOC" link in the upper-right corner of every page if you need documentation for HOC, NEURON's original programming language. HOC may be used in combination with Python: use h.load file to load a HOC library; the functions and classes are then available with an h. prefix. Introduction to Python Displaying results The print command is used to display non-graphical results. It can display fixed text: print ('Hello everyone.') Hello everyone. or the results of a calculation: print (5 * (3 + 2)) 25 Storing results Give values a name to be able to use them later. a = max([1.2, 5.2, 1.7, 3.6]) print (a) 5.2 In Python 2.x, print is a keyword and the parentheses are unnecessary. Using the parentheses allows your code to work with both Python 2.x and 3.x. Don't repeat yourself Lists and for loops To do the same thing to several items, put the items in a list and use a for loop: numbers = [1, 3, 5, 7, 9] for number in numbers: print (number * number) 1 9 25 49 81 Items can be accessed directly using the [] notation; e.g.
    [Show full text]
  • NEST by Example: an Introduction to the Neural Simulation Tool NEST Version 2.6.0
    NEST by Example: An Introduction to the Neural Simulation Tool NEST Version 2.6.0 Marc-Oliver Gewaltig1, Abigail Morrison2, and Hans Ekkehard Plesser3, 2 1Blue Brain Project, Ecole Polytechnique Federale de Lausanne, QI-J, Lausanne 1015, Switzerland 2Institute of Neuroscience and Medicine (INM-6) Functional Neural Circuits Group, J¨ulich Research Center, 52425 J¨ulich, Germany 3Dept of Mathematical Sciences and Technology, Norwegian University of Life Sciences, PO Box 5003, 1432 Aas, Norway Abstract The neural simulation tool NEST can simulate small to very large networks of point-neurons or neurons with a few compartments. In this chapter, we show by example how models are programmed and simulated in NEST. This document is based on a preprint version of Gewaltig et al. (2012) and has been updated for NEST 2.6 (r11744). Updated to 2.6.0 Hans E. Plesser, December 2014 Updated to 2.4.0 Hans E. Plesser, June 2014 Updated to 2.2.2 Hans E. Plesser & Marc-Oliver Gewaltig, December 2012 1 Introduction NEST is a simulator for networks of point neurons, that is, neuron models that collapse the morphol- ogy (geometry) of dendrites, axons, and somata into either a single compartment or a small number of compartments (Gewaltig and Diesmann, 2007). This simplification is useful for questions about the dynamics of large neuronal networks with complex connectivity. In this text, we give a practical intro- duction to neural simulations with NEST. We describe how network models are defined and simulated, how simulations can be run in parallel, using multiple cores or computer clusters, and how parts of a model can be randomized.
    [Show full text]
  • Parallel Range, Segment and Rectangle Queries with Augmented Maps
    Parallel Range, Segment and Rectangle Queries with Augmented Maps Yihan Sun Guy E. Blelloch Carnegie Mellon University Carnegie Mellon University [email protected] [email protected] Abstract The range, segment and rectangle query problems are fundamental problems in computational geometry, and have extensive applications in many domains. Despite the significant theoretical work on these problems, efficient implementations can be complicated. We know of very few practical implementations of the algorithms in parallel, and most implementations do not have tight theoretical bounds. In this paper, we focus on simple and efficient parallel algorithms and implementations for range, segment and rectangle queries, which have tight worst-case bound in theory and good parallel performance in practice. We propose to use a simple framework (the augmented map) to model the problem. Based on the augmented map interface, we develop both multi-level tree structures and sweepline algorithms supporting range, segment and rectangle queries in two dimensions. For the sweepline algorithms, we also propose a parallel paradigm and show corresponding cost bounds. All of our data structures are work-efficient to build in theory (O(n log n) sequential work) and achieve a low parallel depth (polylogarithmic for the multi-level tree structures, and O(n) for sweepline algorithms). The query time is almost linear to the output size. We have implemented all the data structures described in the paper using a parallel augmented map library. Based on the library each data structure only requires about 100 lines of C++ code. We test their performance on large data sets (up to 108 elements) and a machine with 72-cores (144 hyperthreads).
    [Show full text]
  • NEURON and Python
    ORIGINAL RESEARCH ARTICLE published: 28 January 2009 NEUROINFORMATICS doi: 10.3389/neuro.11.001.2009 NEURON and Python Michael L. Hines1, Andrew P. Davison2* and Eilif Muller3 1 Computer Science, Yale University, New Haven, CT, USA 2 Unité de Neurosciences Intégratives et Computationelles, CNRS, Gif sur Yvette, France 3 Laboratory for Computational Neuroscience, Ecole Polytechnique Fédérale de Lausanne, Switzerland Edited by: The NEURON simulation program now allows Python to be used, alone or in combination with Rolf Kötter, Radboud University, NEURON’s traditional Hoc interpreter. Adding Python to NEURON has the immediate benefi t Nijmegen, The Netherlands of making available a very extensive suite of analysis tools written for engineering and science. Reviewed by: Felix Schürmann, Ecole Polytechnique It also catalyzes NEURON software development by offering users a modern programming Fédérale de Lausanne, Switzerland tool that is recognized for its fl exibility and power to create and maintain complex programs. At Volker Steuber, University of the same time, nothing is lost because all existing models written in Hoc, including graphical Hertfordshire, UK user interface tools, continue to work without change and are also available within the Python Arnd Roth, University College London, UK context. An example of the benefi ts of Python availability is the use of the xml module in *Correspondence: implementing NEURON’s Import3D and CellBuild tools to read MorphML and NeuroML model Andrew Davison, UNIC, Bât. 32/33, specifi cations. CNRS, 1 Avenue de la Terrasse, 91198 Keywords: Python, simulation environment, computational neuroscience Gif sur Yvette, France. e-mail: [email protected] INTRODUCTION for the purely numerical issue of how many compartments are The NEURON simulation environment has become widely used used to represent each of the cable sections.
    [Show full text]
  • NEURAL CONNECTIONS: Some You Use, Some You Lose
    NEURAL CONNECTIONS: Some You Use, Some You Lose by JOHN T. BRUER SOURCE: Phi Delta Kappan 81 no4 264-77 D 1999 . The magazine publisher is the copyright holder of this article and it is reproduced with permission. Further reproduction of this article in violation of the copyright is prohibited JOHN T. BRUER is president of the James S. McDonnell Foundation, St. Louis. This article is adapted from his new book, The Myth of the First Three Years (Free Press, 1999), and is reprinted by arrangement with The Free Press, a division of Simon Schuster Inc. ©1999, John T. Bruer . OVER 20 YEARS AGO, neuroscientists discovered that humans and other animals experience a rapid increase in brain connectivity -- an exuberant burst of synapse formation -- early in development. They have studied this process most carefully in the brain's outer layer, or cortex, which is essentially our gray matter. In these studies, neuroscientists have documented that over our life spans the number of synapses per unit area or unit volume of cortical tissue changes, as does the number of synapses per neuron. Neuroscientists refer to the number of synapses per unit of cortical tissue as the brain's synaptic density. Over our lifetimes, our brain's synaptic density changes in an interesting, patterned way. This pattern of synaptic change and what it might mean is the first neurobiological strand of the Myth of the First Three Years. (The second strand of the Myth deals with the notion of critical periods, and the third takes up the matter of enriched, or complex, environments.) Popular discussions of the new brain science trade heavily on what happens to synapses during infancy and childhood.
    [Show full text]
  • Handwritten Digit Classication Using 8-Bit Floating Point Based Convolutional Neural Networks
    Downloaded from orbit.dtu.dk on: Apr 10, 2018 Handwritten Digit Classication using 8-bit Floating Point based Convolutional Neural Networks Gallus, Michal; Nannarelli, Alberto Publication date: 2018 Document Version Publisher's PDF, also known as Version of record Link back to DTU Orbit Citation (APA): Gallus, M., & Nannarelli, A. (2018). Handwritten Digit Classication using 8-bit Floating Point based Convolutional Neural Networks. DTU Compute. (DTU Compute Technical Report-2018, Vol. 01). General rights Copyright and moral rights for the publications made accessible in the public portal are retained by the authors and/or other copyright owners and it is a condition of accessing publications that users recognise and abide by the legal requirements associated with these rights. • Users may download and print one copy of any publication from the public portal for the purpose of private study or research. • You may not further distribute the material or use it for any profit-making activity or commercial gain • You may freely distribute the URL identifying the publication in the public portal If you believe that this document breaches copyright please contact us providing details, and we will remove access to the work immediately and investigate your claim. Handwritten Digit Classification using 8-bit Floating Point based Convolutional Neural Networks Michal Gallus and Alberto Nannarelli (supervisor) Danmarks Tekniske Universitet Lyngby, Denmark [email protected] Abstract—Training of deep neural networks is often con- In order to address this problem, this paper proposes usage strained by the available memory and computational power. of 8-bit floating point instead of single precision floating point This often causes it to run for weeks even when the underlying which allows to save 75% space for all trainable parameters, platform is employed with multiple GPUs.
    [Show full text]
  • High Dynamic Range Video
    High Dynamic Range Video Karol Myszkowski, Rafał Mantiuk, Grzegorz Krawczyk Contents 1 Introduction 5 1.1 Low vs. High Dynamic Range Imaging . 5 1.2 Device- and Scene-referred Image Representations . ...... 7 1.3 HDRRevolution ............................ 9 1.4 OrganizationoftheBook . 10 1.4.1 WhyHDRVideo? ....................... 11 1.4.2 ChapterOverview ....................... 12 2 Representation of an HDR Image 13 2.1 Light................................... 13 2.2 Color .................................. 15 2.3 DynamicRange............................. 18 3 HDR Image and Video Acquisition 21 3.1 Capture Techniques Capable of HDR . 21 3.1.1 Temporal Exposure Change . 22 3.1.2 Spatial Exposure Change . 23 3.1.3 MultipleSensorswithBeamSplitters . 24 3.1.4 SolidStateSensors . 24 3.2 Photometric Calibration of HDR Cameras . 25 3.2.1 Camera Response to Light . 25 3.2.2 Mathematical Framework for Response Estimation . 26 3.2.3 Procedure for Photometric Calibration . 29 3.2.4 Example Calibration of HDR Video Cameras . 30 3.2.5 Quality of Luminance Measurement . 33 3.2.6 Alternative Response Estimation Methods . 33 3.2.7 Discussion ........................... 34 4 HDR Image Quality 39 4.1 VisualMetricClassification. 39 4.2 A Visual Difference Predictor for HDR Images . 41 4.2.1 Implementation......................... 43 5 HDR Image, Video and Texture Compression 45 1 2 CONTENTS 5.1 HDR Pixel Formats and Color Spaces . 46 5.1.1 Minifloat: 16-bit Floating Point Numbers . 47 5.1.2 RGBE: Common Exponent . 47 5.1.3 LogLuv: Logarithmic encoding . 48 5.1.4 RGB Scale: low-complexity RGBE coding . 49 5.1.5 LogYuv: low-complexity LogLuv . 50 5.1.6 JND steps: Perceptually uniform encoding .
    [Show full text]
  • The Action Potential
    See discussions, stats, and author profiles for this publication at: http://www.researchgate.net/publication/6316219 The action potential ARTICLE in PRACTICAL NEUROLOGY · JULY 2007 Source: PubMed CITATIONS READS 16 64 2 AUTHORS, INCLUDING: Mark W Barnett The University of Edinburgh 21 PUBLICATIONS 661 CITATIONS SEE PROFILE Available from: Mark W Barnett Retrieved on: 24 October 2015 192 Practical Neurology HOW TO UNDERSTAND IT Pract Neurol 2007; 7: 192–197 The action potential Mark W Barnett, Philip M Larkman t is over 60 years since Hodgkin and called ion channels that form the permeation Huxley1 made the first direct recording of pathways across the neuronal membrane. the electrical changes across the neuro- Although the first electrophysiological nal membrane that mediate the action recordings from individual ion channels were I 2 potential. Using an electrode placed inside a not made until the mid 1970s, Hodgkin and squid giant axon they were able to measure a Huxley predicted many of the properties now transmembrane potential of around 260 mV known to be key components of their inside relative to outside, under resting function: ion selectivity, the electrical basis conditions (this is called the resting mem- of voltage-sensitivity and, importantly, a brane potential). The action potential is a mechanism for quickly closing down the transient (,1 millisecond) reversal in the permeability pathways to ensure that the polarity of this transmembrane potential action potential only moves along the axon in which then moves from its point
    [Show full text]
  • Supporting Technology for Augmented Reality Game-Based Learning
    DOCTORAL THESIS Supporting Technology for Augmented Reality Game-Based Learning Hendrys Fabián Tobar Muñoz 2017 Doctorate in Technology Supervised by: PhD. Ramon Fabregat PhD. Silvia Baldiris Presented in partial fulfillment of the requirements for a doctoral degree from the University of Girona The research reported in this thesis was partially sponsored by: The financial support by COLCIENCIAS – Colombia’s Administrative Department of Science Technology and Innovation within the program of doctoral scholarships. The research reported in this thesis was carried out as part of the following projects: ARreLS project, funded by Spanish Economy and Competitiveness Ministry(TIN2011-23930) Open Co-Creation Project, funded by the Spanish Economy and Competitiveness Ministry (TIN2014-53082-R) The research reports in this thesis was developed within the research lines of the Communications and Distributed Systems research group (BCDS, ref GRCT40) which is part of the DURSI consolidated research group COMUNICACIONS I SISTEMES INTEL·LIGENTS. © Hendrys Fabián Tobar Muñoz, Girona, Catalonia, Spain, 2017 All Rights Reserved. No part of this book may be reproduced in any form by any electronic or mechanical means (including photocopying, recording, or information storage and retrieval) without written permission granted by the author. Gracias a Dios Todopoderoso y a Su Divina Providencia. Este trabajo está dedicado a mi Reina Omaira que es lo más hermoso de este mundo. ______________________________________________________________________________________________________________________ Thanks to God Almighty and His Divine Providence. This work is dedicated to my Queen Omaira who is the most beautiful of this world (I swear, this sounds better in Spanish) ACKNOWLEDGEMENTS I have always liked games and digital technologies and I have always believed in their educational potential.
    [Show full text]
  • On the Automated Derivation of Domain-Specific UML Profiles
    Schriften aus der Fakultät Wirtschaftsinformatik und 37 Angewandte Informatik der Otto-Friedrich-Universität Bamberg On the Automated Derivation of Domain-Specifc UML Profles Alexander Kraas 37 Schriften aus der Fakultät Wirtschaftsinformatik und Angewandte Informatik der Otto-Friedrich- Universität Bamberg Contributions of the Faculty Information Systems and Applied Computer Sciences of the Otto-Friedrich-University Bamberg Schriften aus der Fakultät Wirtschaftsinformatik und Angewandte Informatik der Otto-Friedrich- Universität Bamberg Contributions of the Faculty Information Systems and Applied Computer Sciences of the Otto-Friedrich-University Bamberg Band 37 2019 On the Automated Derivation of Domain-Specifc UML Profles Alexander Kraas 2019 Bibliographische Information der Deutschen Nationalbibliothek Die Deutsche Nationalbibliothek verzeichnet diese Publikation in der Deutschen Nationalbibliographie; detaillierte bibliographische Informationen sind im Internet über http://dnb.d-nb.de/ abrufbar. Diese Arbeit hat der Fakultät Wirtschaftsinformatik und Angewandte Informatik der Otto-Friedrich-Universität Bamberg als Dissertation vorgelegen. 1. Gutachter: Prof. Dr. Gerald Lüttgen, Otto-Friedrich-University Bamberg, Germany 2. Gutachter: Prof. Dr. Richard Paige, McMaster University, Canada Tag der mündlichen Prüfung: 13.05.2019 Dieses Werk ist als freie Onlineversion über den Publikationsserver (OPUS; http://www. opus-bayern.de/uni-bamberg/) der Universität Bamberg erreichbar. Das Werk – ausge- nommen Cover, Zitate und Abbildungen –
    [Show full text]