Red Hat Integration 2020-Q2 Data Virtualization Tutorial

Total Page:16

File Type:pdf, Size:1020Kb

Red Hat Integration 2020-Q2 Data Virtualization Tutorial Red Hat Integration 2020-Q2 Data Virtualization Tutorial TECHNOLOGY PREVIEW - Data Virtualization Tutorial Last Updated: 2020-07-11 Red Hat Integration 2020-Q2 Data Virtualization Tutorial TECHNOLOGY PREVIEW - Data Virtualization Tutorial Legal Notice Copyright © 2020 Red Hat, Inc. The text of and illustrations in this document are licensed by Red Hat under a Creative Commons Attribution–Share Alike 3.0 Unported license ("CC-BY-SA"). An explanation of CC-BY-SA is available at http://creativecommons.org/licenses/by-sa/3.0/ . In accordance with CC-BY-SA, if you distribute this document or an adaptation of it, you must provide the URL for the original version. Red Hat, as the licensor of this document, waives the right to enforce, and agrees not to assert, Section 4d of CC-BY-SA to the fullest extent permitted by applicable law. Red Hat, Red Hat Enterprise Linux, the Shadowman logo, the Red Hat logo, JBoss, OpenShift, Fedora, the Infinity logo, and RHCE are trademarks of Red Hat, Inc., registered in the United States and other countries. Linux ® is the registered trademark of Linus Torvalds in the United States and other countries. Java ® is a registered trademark of Oracle and/or its affiliates. XFS ® is a trademark of Silicon Graphics International Corp. or its subsidiaries in the United States and/or other countries. MySQL ® is a registered trademark of MySQL AB in the United States, the European Union and other countries. Node.js ® is an official trademark of Joyent. Red Hat is not formally related to or endorsed by the official Joyent Node.js open source or commercial project. The OpenStack ® Word Mark and OpenStack logo are either registered trademarks/service marks or trademarks/service marks of the OpenStack Foundation, in the United States and other countries and are used with the OpenStack Foundation's permission. We are not affiliated with, endorsed or sponsored by the OpenStack Foundation, or the OpenStack community. All other trademarks are the property of their respective owners. Abstract Combine data from multiple sources so that applications can connect to a single, virtual data model Table of Contents Table of Contents .C . H. .A . P. .T .E . R. 1.. .O . .V . E. .R .V . I. E. W. .O . .F . T. .H . E. D. .A . T. .A . .V . I.R . T. U. A. .L .I Z. .A . T. .I O. .N . T. U. .T . O. R. I. A. .L . 3. .C . H. .A . P. .T .E . R. 2. G. E. T. .T . I.N . G. .S .T . A. .R . T. .E .D . W. I.T .H . D. .A . T. A. V. .I .R .T . U. .A . L. .I Z. .A . T. I. O. .N . .4 . .C . H. .A . P. .T .E . R. 3. I. N. .T . R. .O . D. .U . C. .T . I.O . .N . .T . O. D. .A . T. .A . .V . I.R . T. .U . A. .L .I Z. .A . T. .I O. .N . 6. .C . H. .A . P. .T .E . R. 4. .S .E . T. .T . I.N . G. U. .P . .T . H. .E . .E . N. .V . I.R . O. .N . .M . E. .N . T. 7. 4.1. CLONING THE TUTORIAL RESOURCES 7 4.2. CREATING THE SOURCE DATABASE 7 .C . H. .A . P. .T .E . R. 5. A. .D . .D . I.N . G. .T .H . E. D. .A . T. .A . .V . I.R . T. .U . A. .L .I .Z .A . T. .I O. .N . O. .P . E. .R . A. .T .O . .R . .T .O . T. .H . E. .P . R. .O . .J .E . C. .T . 1.0 . 5.1. CREATING A DOCKER-REGISTRY PULL SECRET 10 5.2. INSTALLING THE DATA VIRTUALIZATION OPERATOR IN YOUR PROJECT 11 5.3. LINKING THE PULL SECRET TO THE DATA VIRTUALIZATION OPERATOR SERVICE ACCOUNT 12 .C . H. .A . P. .T .E . R. 6. .B . U. .I L. .D . I.N . G. T. .H . E. V. .I R. .T . U. .A . L. D . .A . T. A. .B . A. .S . E. 1.4 . 6.1. VIRTUAL DATABASE CUSTOM RESOURCES 14 6.1.1. Data source configuration 14 6.1.2. DDL for defining the virtual database 15 6.1.2.1. Virtual database creation 16 6.1.2.2. Translator definition 16 6.1.2.3. External data source definitions 16 6.1.2.4. Schema creation 17 6.1.2.5. Metadata import 17 6.1.2.6. Virtual schema definition 17 6.2. COMPLETED VIRTUAL DATABASE CUSTOM RESOURCE FILE 19 .C . H. .A . P. .T .E . R. 7. D. E. P. .L . O. .Y . I.N . .G . .T . H. .E . .V . I.R . T. U. .A . L. D. .A . T. .A . B. .A . S. .E . 2. .1 . .C . H. .A . P. .T .E . R. 8. .C . O. .N . N. E. C. .T . I.N . .G . .C . L. .I E. .N . T. .S . T. .O . .T . H. .E . .V . I.R . T. .U . A. .L . D. A. .T .A . B. .A . S. .E . 2. 2. 8.1. CONNECTING AN INTERNAL JDBC CLIENT 22 8.1.1. Installing SQLLine 22 8.1.2. Connecting SQLLine to the Portfolio virtual database 23 8.2. CONNECTING TO THE VIRTUAL DATABASE FROM AN EXTERNAL JDBC CLIENT 24 8.2.1. Configuring an OpenShift load balancer service to enable external JDBC clients to access the virtual database 24 8.2.2. Enabling external JDBC client access through port forwarding 25 8.2.3. Installing the SQuirreL JDBC client 26 8.2.4. Configuring SQuirreL to connect to the Portfolio virtual database 27 8.2.5. Querying the Portfolio virtual database from the SQuirreL SQL client 28 8.3. SAMPLE QUERIES 28 8.4. ACCESS THE VIRTUAL DATABASE THROUGH THE ODATA API 29 1 Red Hat Integration 2020-Q2 Data Virtualization Tutorial 2 CHAPTER 1. OVERVIEW OF THE DATA VIRTUALIZATION TUTORIAL CHAPTER 1. OVERVIEW OF THE DATA VIRTUALIZATION TUTORIAL This tutorial demonstrates how to use Red Hat Integration data virtualization to create a customer portfolio virtual database. The virtual database that you create integrates data from the following two sources: A postgreSQL accounts database Stores customer account information, and data about the stock holdings for each customer. A web-based REST service Provides live stock market price data. By integrating data from these two sources, the Portfolio database calculates the value of individual customer portfolios based on current stock prices. After we deploy the virtual database, we’ll submit queries to it to demonstrate how data from both of sources is combined. Figure 1.1. Architecture of the Portfolio virtual database in this tutorial 3 Red Hat Integration 2020-Q2 Data Virtualization Tutorial CHAPTER 2. GETTING STARTED WITH DATA VIRTUALIZATION In this tutorial we demonstrate how to create and use a "Portfolio" virtual database by completing the following tasks: 1. Setting up our environment. 2. Adding a Data Virtualization Operator to OpenShift. 3. Creating a postgreSQL accounts database with a simple sample database to serve as our data source. 4. Creating a custom resource (CR) that defines a Portfolio virtual database. The CR specifies how to integrate data from our postgreSQL database and a REST service. 5. Running the Operator to deploy the virtual database. 6. Demonstrating how JDBC and OData clients can access and query the virtual database. NOTE ODBC access is also available, but information about how to enable ODBC access is beyond the scope of this tutorial. Time 30 minutes to 1 hour Skill Beginner IMPORTANT Data virtualization is a Technology Preview feature only. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process. For more information about the support scope of Red Hat Technology Preview features, see https://access.redhat.com/support/offerings/techpreview/. Prerequisites Before you can use data virtualization to create a virtual database, you must have: Access to an OpenShift 4.4 cluster with cluster-admin privileges. This can be an enterprise deployment of OpenShift (OpenShift Container Platform, OpenShift Dedicated, or OpenShift Online), or a local installation that uses Red Hat Code Ready Containers (CRC). For more information about CRC, see the article on Red Hat Developer. A version of the OpenShift CLI (oc) binary that matches the version of the OpenShift server. Working knowledge of SQL. Working knowledge of Openshift and the OpenShift Operator model. 4 CHAPTER 2. GETTING STARTED WITH DATA VIRTUALIZATION Working knowledge of git and GitHub. 5 Red Hat Integration 2020-Q2 Data Virtualization Tutorial CHAPTER 3. INTRODUCTION TO DATA VIRTUALIZATION A virtual database enables you to aggregate data from one or more external data sources and apply a custom schema to the data. By applying a customized logical schema to the aggregated source data, you can curate the data in a way that makes it easy for your applications to consume. The virtual view exposes just the data that you want from each data source, down to the level of specific tables, columns or procedures. Query processing logic in the virtual database enables users to access and join data in different formats from across data sources. The virtual database provides an abstraction layer that shields client applications from the details of the physical data sources. Data consumers don’t have figure out how to connect to the host sources. They also don’t have to worry about how a view combines data from various sources, or how to configure translators to normalize data into a usable format. Data and access where you want it Data virtualization acts as a logical data warehouse, one that relies on metadata to make data available to client applications.
Recommended publications
  • Sql Plus Download Mac
    Sql plus download mac LINK TO DOWNLOAD oracle sql plus free download - Insight Developer for Oracle, SQLite Database, Orac, and many more programs. SQL*Plus Instant Client can be installed in two ways: Download the packages from the Oracle Technology Network (OTN). Copy the same files that are in the packages from an Oracle Database 10 g Client Administrator installation. Both the SQL*Plus and OCI packages must be from the same Oracle Database version, for example, SQL*PLus on Mac. December 1st, Goto comments Leave a comment. I would think installing SQL*Plus on the Mac would be point, click download, point click, install bam it works. Nah. It did install mostly straight forward on my old Mac. Got a new Mac and no dice. Jul 30, · Download the (free) Docker Community Edition for Mac (unless you’ve already got it installed on your system). This will enable you to run SQL Server from within a Docker container. To download, visit the Docker CE for Mac download page and click Get Docker. Find "Macintosh OSX" and download both the Basic and SQL*Plus packages from the Instant Client downloads page. Move the contents of both into a single convenient folder (I used /Applications/Application_folders/instantclient). Add the location to your PATH variable, and also set DYLD_LIBRARY_PATH to point to it; for example. That could affect running SQL*Plus from a shell script, for example. There are workarounds for the 11g instant client. The installation notes at the bottom of the download page have changed since I last did this, and it now says to hard link the library files to the user's ~/lib directory to avoid that issue.
    [Show full text]
  • Squirrel SQL Client
    Pagina: 1 di 22 Versione: 01.09 Data agg.: 08/02/2021 RM Squirrel SQL client guida rapida Autori: Marco Riva Revisori: g:\installazioni\squirrel sql client 4.1.0\squirrel sql client - guida rapida.docx Squirrel SQL client Pagina: 2 di 22 guida rapida Versione: 01.09 Data agg.: 08/02/2021 RM Sommario 1. Introduzione ........................................ 3 1.1. Fast startup .......................................... 3 2. Installazione ........................................ 4 2.1. Requisiti ............................................... 4 2.2. Download ............................................. 4 2.3. Installazione .......................................... 4 2.4. Aggiornamento ...................................... 4 3. Overview ............................................ 5 4. Configurazione ..................................... 5 4.1. Configurazione driver per DB2 for i ........... 5 4.2. Configurazione alias ............................... 7 5. Utilizzo .............................................. 12 5.1. Connessione ad un database .................. 12 5.2. Maggiori informazioni sull’area object ...... 14 5.2.1. Filtro e ricerca oggetti ..................... 14 5.2.2. Scheda content .............................. 15 5.2.3. Esportazione contenuti.................... 15 5.2.4. Modifica diretta in tabella ................ 16 5.3. Editor SQL .......................................... 17 5.3.1. Scorciatoie da tastiera .................... 17 5.3.2. Regole sintassi .............................. 17 5.3.3. Barra dei pulsanti ..........................
    [Show full text]
  • Presto: the Definitive Guide
    Presto The Definitive Guide SQL at Any Scale, on Any Storage, in Any Environment Compliments of Matt Fuller, Manfred Moser & Martin Traverso Virtual Book Tour Starburst presents Presto: The Definitive Guide Register Now! Starburst is hosting a virtual book tour series where attendees will: Meet the authors: • Meet the authors from the comfort of your own home Matt Fuller • Meet the Presto creators and participate in an Ask Me Anything (AMA) session with the book Manfred Moser authors + Presto creators • Meet special guest speakers from Martin your favorite podcasts who will Traverso moderate the AMA Register here to save your spot. Praise for Presto: The Definitive Guide This book provides a great introduction to Presto and teaches you everything you need to know to start your successful usage of Presto. —Dain Sundstrom and David Phillips, Creators of the Presto Projects and Founders of the Presto Software Foundation Presto plays a key role in enabling analysis at Pinterest. This book covers the Presto essentials, from use cases through how to run Presto at massive scale. —Ashish Kumar Singh, Tech Lead, Bigdata Query Processing Platform, Pinterest Presto has set the bar in both community-building and technical excellence for lightning- fast analytical processing on stored data in modern cloud architectures. This book is a must-read for companies looking to modernize their analytics stack. —Jay Kreps, Cocreator of Apache Kafka, Cofounder and CEO of Confluent Presto has saved us all—both in academia and industry—countless hours of work, allowing us all to avoid having to write code to manage distributed query processing.
    [Show full text]
  • Downloads/PDF for Archive.Pdf
    European Space Agency Symposium “Ensuring Long-Term Preservation and Adding Value to Scientific and Technical Data”, 5 - 7 October 2004, Frascati, Italy Providing Authentic Long-term Archival Access to Complex Relational Data Stephan Heuscher a, Stephan J¨armann b, Peter Keller-Marxer c, and Frank M¨ohle d Swiss Federal Archives, Archivstrasse 24, CH-3003 Bern, Switzerland [email protected], [email protected], [email protected], [email protected] (Dated: July 19, 2004) We discuss long-term preservation of and access to relational databases. The focus is on national archives and science data archives which have to ingest and integrate data from a broad spectrum of vendor-specific relational database management systems (RDBMS). Furthermore, we present our solution SIARD which analyzes and extracts data and data logic from almost any RDBMS. It enables, to a reasonable level of authenticity, complete detachment of databases from their vendor-specific environment. The user can add archival descriptive metadata according to a customizable schema. A SIARD database archive integrates data, data logic, technical metadata, and archival descriptive information in one archival information package, independent of any specific software and hardware, based upon plain text files and the standardized languages SQL and XML. For usage purposes, a SIARD archive can be reloaded into any current or future RDBMS which supports standard SQL. In addition, SIARD contains a client that enables ‘on demand’ reload of archives into a target RDBMS, and multi-user remote access for querying and browsing the data together with its technical and descriptive metadata in one graphical user interface.
    [Show full text]
  • Vibe 4.0.8 Installation Guide
    Vibe 4.0.8 Installation Guide March 2021 Legal Notice Copyright © 2018 - 2021 Micro Focus or one of its affiliates. The only warranties for products and services of Micro Focus and its affiliates and licensors (“Micro Focus”) are as may be set forth in the express warranty statements accompanying such products and services. Nothing herein should be construed as constituting an additional warranty. Micro Focus shall not be liable for technical or editorial errors or omissions contained herein. The information contained herein is subject to change without notice. 2 Contents About This Guide 11 Part I Overview of Micro Focus Vibe 13 1 Vibe’s User Types 15 2 Vibe’s Service Components 17 3 Topology of Vibe Deployment Options 19 Part II System Requirements and Support 21 4 Vibe System Requirements 23 Server Hardware Requirements. .23 Server Operating System Requirements . .23 Database Server Requirements . .24 Directory Service (LDAP) Requirements . .25 Disk Space Requirements . .26 Other Requirements . .26 5 Vibe User Platform Support 27 Browser Support . .27 Office Add-In Support. .27 Office Application WebDAV Support . .28 Collaboration Client Support . .28 Desktop Support . .28 Mobile Support . .29 App Availability . .29 Browser Requirements . .29 File Viewer Support. .29 6 Environment Support 31 IPV6 Support . .31 Clustering Support . .31 Virtualization Support. .31 Single Sign-On Support. .31 Linux File System Support . .32 Contents 3 Part III Single-server (Basic) Installation 33 7 Single-server Installation Planning Worksheet 35 8 Planning a Basic (Single-server) Vibe Installation 41 What Is a Basic Vibe Installation? . .41 Planning the Operating Environment of Your Vibe Server.
    [Show full text]
  • Bilag Til CV Tekniske Kompetencer Søren Lund [Skjult I Online Version
    Bilag til CV Tekniske kompetencer Søren Lund Jeg er en alsidig it-konsulent [skjult i online version] med mere end 20 ars˚ profes- [skjult i online version] sionel erfaring. Jeg har en [skjult i online version] bred vifte af kompetencer, og k [email protected] kan lide nye udfordringer. Nyeste udgave kan hentes her: https: // hub. 369. dk/ resume/ kompetencer. pdf Dato og tid for seneste ændring af denne udgave: 2019-09-07 09:38 Herunder er en gennemgang af mine tekniske kompetencer. For hver kompetence har jeg angivet et niveau (Meget højt, Højt, Godt kendskab, Noget kendskab), hvorn˚arjeg sidst har brugt den p˚agældendekompetence, samt hvor mange ˚arsprofessionel erfaring jeg har. Enkelte steder har jeg desuden angivet et ˚arstali parentes, det angiver hvor mange ˚arjeg i alt har erfaring, f.eks. fra studiet eller privat. Arbejdsomr˚ader/-processer Niveau Sidst brugt Ar˚ Analyse Højt 2019 23 Databasedesign Højt 2018 17 Design Højt 2019 23 Design af brugergrænseflader Højt 2014 15 Implementering Meget højt 2019 23 Kvalitetssikring Meget højt 2019 23 Programmering Højt 2019 23 Sikkerhed Godt kendskab 2019 17 Systemadministration Højt 2019 23 Udrulning Højt 2019 23 Webteknologi Meget højt 2019 21 1 Platforme Niveau Sidst brugt Ar˚ ASP Godt kendskab 2005 6 Apache FOP Godt kendskab 2016 4 Apache HTTP Server Højt 2017 17 Apache Tomcat Højt 2018 12 Amazon Web Services (AWS) Godt kendskab 2019 1 CouchDB Godt kendskab 2011 1 Docker Noget kendskab 2019 1 Elastic Path eCommerce platform Noget kendskab 2006 1 Hadoop Noget kendskab 2013 <1 HP-UX (bruger/server)
    [Show full text]
  • Modeling User Behavior by Integrating AQ Learning with a Database: Initial Results
    Modeling User Behavior by Integrating AQ Learning with a Database: Initial Results Guido Cervone and Ryszard S. Michalski* Machine Learning and Inference Laboratory School Computational Sciences George Mason University Fairfax, VA, 22030 {gcervone, michalski}@gmu.edu *Also with the Institute of Computer Science, Polish Academy of Sciences, Warsaw, Poland Abstract: The paper describes recent results from developing and testing LUS methodology for user modeling. LUS employs AQ learning for automatically creating user models from datasets representing activities of computer users. The datasets are stored in a relational database and employed in the learning process through an SQL-style command that automatically executes the AQ20 rule learning program and generates user models. The models are in the form of attributional rulesets that are more expressive than conventional decision rules, and are easy to interpret and understand. Early experimental results from the testing of the LUS method gave highly encouraging results. Keywords: User modeling, Computer intrusion detection, Machine learning, AQ learning, Inductive databases 1 Introduction The rapidly growing global connectivity of computer systems creates a great need for effective methods that are able to detect unauthorized use of computers. Standard methods for assuring computer security, such as passwords, gateways, and firewalls not always provide sufficient protection from unauthorized accesses. Intruders typically exploit holes in the operating system or crack password files to gain access to the computer system and masquerade as legitimate users. As a result, detection of a sophisticated intruder is increasingly difficult, especially when there are many computer users or the intruder is an insider. The approach discussed in this paper, called Learning User Style (LUS), applies symbolic learning, specifically AQ learning, to induce typical patterns of interactions between individual users and computers.
    [Show full text]
  • Oracle Jdbc Url Schema
    Oracle Jdbc Url Schema radiologistStephanus besomsputtings tooGermanically? insipidly? Ambrosio invites gregariously. Darius remains idiographic: she lamb her Configuring an Oracle database Alfresco Documentation. Method must exercise you are proprietary, oracle jdbc url schema names, one method can also known as vendor. This schema information for oracle database using enterprise java. Jira database service name of available from python will notice there are not exist at all stored procedures, of milliseconds that are no password? If another customer engagement and oracle jdbc url and so you retrieve an unsecure phoenix driver, url below and share connections. Setup first time SonarQube SonarSource Community. If enabled oracle jdbc url schema is listening port. Select the driver class appropriate title your JDBC database connection. Open in oracle jdbc url schema. Save settings and schemas to protect from human errors are sent to. With references or view. TAF requires an acknowledgement from the application that school failure has occurred through a rollback command. The schema is controlled by Appian and present not install any tables created by. This url to oracle is setup properly with underscores should be applied to be different schemas to connect to make sure if my blog. Therefore, put other settings. This oxygen has been undeleted. Make a jdbc urls which of schemas. Jdbc connect schema is, and which genesys info mart database service name of oracle database using this section in that is placed before users. Database Setup through Flyway. From a maintenance point of black it is beneficial to poison multiple tablespaces for different people of objects. The above entry must be reading below the connection.
    [Show full text]
  • Tips Table of Contents 1
    ART Tips Table of Contents 1. Using the jTDS driver . 1 2. Migrating the ART Database . 1 2.1. Migrating between the same RDBMS . 1 2.2. Migrating between different RDBMSs . 1 2.2.1. Using Pentaho Data Integration . 1 2.2.2. Migrating the demo database . 2 3. Upgrading a HSQLDB database . 2 4. Deployment on JBoss 7 . 3 5. Using custom fonts in JasperReports . 3 6. Strange characters with charts . 4 - i - 1. Using the jTDS driver If connecting to a named instance e.g. if you are using SQL Server 2008 Express, use a JDBC url like the following jdbc:jtds:sqlserver://localhost/mydatabase;instance=SQLEXPRESS 2. Migrating the ART Database ART doesn’t provide any in-built tools to migrate the repository from one database to another. The following are suggested techniques for such a migration. 2.1. Migrating between the same RDBMS To migrate between the same RDBMS e.g. from a MySQL database to another MySQL database, use the backup/restore tools provided for your RDBMS 2.2. Migrating between different RDBMSs To migrate between different RDBMSs e.g. from a PostgreSQL database to a MySQL database, use an ETL tool e.g. Pentaho Data Integration 2.2.1. Using Pentaho Data Integration These steps illustrate the steps to migrate an ART repository database on PostgreSQL to MySQL. Create a new database on your MySQL server to serve as the destination. Run the art-tables.sql and quartz\tables_mysql.sql scripts on this database to create a new, empty ART repository Run the spoon.bat file in the PDI_HOME\data-integration directory From the Tools | Wizards | Create database connection menu, create connections for your source and destination database Select the Tools | Wizards | Copy tables menu Select the source and destination databases In the available items list, select all the tables except the qrtz_xx tables.
    [Show full text]
  • Rapport Projet Ahmed
    ÉCOLE DE TECHNOLOGIE SUPÉRIEURE UNIVERSITÉ DU QUÉBEC RAPPORT DE PROJET PRÉSENTÉ À L’ÉCOLE DE TECHNOLOGIE SUPÉRIEURE COMME EXIGENCE PARTIELLE À L’OBTENTION DE LA MAÎTRISE EN GÉNIE LOGICIEL M.Ing. PAR Ahmed SEDJAI ÉTUDES D’OUTILS DE CONTRÔLE DE LA QUALITE LOGICIEL (LOGICIEL LIBRE) POUR SUPPORTER LES NORMES DE GÉNIE LOGICIEL MONTRÉAL, LE 31 AOÛT 2013 Ahmed Sedjai, 2013 ©Tous droits réservés Cette licence signifie qu’il est interdit de reproduire, d’enregistrer ou de diffuser en tout ou en partie, le présent document. Le lecteur qui désire imprimer ou conserver sur un autre media une partie importante de ce document, doit obligatoirement en demander l’autorisation à l’auteur. Cette licence Creative Commons signifie qu’il est permis de diffuser, d’imprimer ou de sauvegarder sur un autre support une partie ou la totalité de cette œuvre à condition de mentionner l’auteur, que ces utilisations soient faites à des fins non commerciales et que le contenu de l’œuvre n’ait pas été modifié. REMERCIEMENTS Je tiens à remercier particulièrement mon directeur de projet, Monsieur Alain APRIL, pour ses recommandations et ses conseils, sans lesquelles je ne pourrai arriver à un tel résultat. Aussi, d’avoir dirigé et consciencieusement relu mon rapport, de m’avoir permis de participer à un tel projet et tout le soutien accordé. À mes parents, partis trop tôt pour me lire. Je remercie toute ma famille, pour leurs encouragements et leur soutient tout au long de mes études. À ma femme, à mes enfants. Aussi, sans oublier tous mes amis. Merci. ÉTUDES D’OUTILS DE CONTRÔLE DE LA QUALITE LOGICIEL (LOGICIEL LIBRE) POUR SUPPORTER LES NORMES DE GÉNIE LOGICIEL Ahmed SEDJAI RÉSUMÉ Notre projet vise à identifier des logiciels libres qui supportent certaines activités de génie logiciel.
    [Show full text]
  • Getting Started Using the Squirrel SQL Client
    Getting Started using the SQuirreL SQL Client The SQuirreL SQL Client is a graphical program written in the Java programming language that will allow you to view the structure of a JDBC-compliant database, browse the data in tables, issue SQL commands and much more. This document will demonstrate how install the client and configure it to connect to and work with an Apache Derby database. Derby is a full-featured, open source relational database management system (RDBMS) that is based on Java technology and SQL (The Apache Software Foundation 2010). The Derby database was selected because it can be setup very easily. The steps for configuring the SQuirreL SQL Client to work with any other database management system will be very similar. Prerequisites The SQuirreL SQL Client is a Java application, and as such, it requires a Java 6 runtime environment (JRE) which can be downloaded from Oracle here: http://www.java.com/en/download/index.jsp. Follow the instructions downloading and installing Java 6 using the executable installer. Earlier versions of the JRE (for example Java 5) are not compatible with the latest versions of the SQuirreL SQL Client. Apple‟s current version of Mac OS X (10.6/Snow Leopard) comes with Java 6 pre- installed. However, Apple and Oracle have made arrangements for Oracle to ship updates to the JRE for Mac OS X in future versions of that operating system, so for newer releases of Mac OS X, check Oracle‟s website for updates. Once the Java 6 runtime environment is installed, the install jar for the SQuirreL SQL Client can be found here: http://www.squirrelsql.org/#installation.
    [Show full text]
  • COPYRIGHTED MATERIAL61 Process Actions And, 85–89 Pentaho Administrative Console
    Index A Active@ ISO Burner, 22 absolute positioning, PRD, 398 Ad Hoc Report Wizard, 373–375 access control Add Cube button, 466–467 metadata layer refining privileges, 349 Add Job process action, Scheduler, schema design and, 483 418–419 accumulating periodic snapshot fact table, Add Levels, 474–476 149–150 Add Parameter function, PRD, 386–389 action definitions, 68 Add Sequence step, PDI action sequence editor, 80–82 loading date dimension, 268–269 Action Sequence Wizard, 80, 86 loading demography dimension, 283, 284 action sequences additive measures, 150 Addresses tab page, Mail job, 290 adding as templates for Design Studio, 88 Ad-Hoc Report component, 192 creating with PDS. See PDS (Pentaho admin user Design Studio) creating slave servers, 340 Customers per Website pie chart, managing PDI repository accounts, 548–551 326–327 executed by solution engine, 68 PDI repository, 324 executing in background, 422–423 administration-console, 38, 44 functionality of, 78 administrative tasks inputs for, 83–85 data sources, 60–61 for location data, 559–561 managing schedules and subscriptions, outputs for, 85 COPYRIGHTED MATERIAL61 process actions and, 85–89 Pentaho Administrative Console. See programming Scheduler with, 412, PAC (Pentaho Administrative 417–420 Console) running jobs within, 335 user management, 58–60 solution repository containing, 68 Administrator profile, PDI repository, 327 subscribing to, 423–426 Advanced category, Database Connection using transformations in, 334–336 dialog, 250 actions, process, 85–89 Advisor button, PAD, 501 Active Directory (AD), and EE single AgeandAgeSequencestep,demography sign-on, 77 dimensions, 282 571 572 Index ■ A Age Group step, demography dimensions, ARFF (Attribute Relation File Format), 282, 284–285 511, 519 Aggregate Designer, 130, 442 AS.
    [Show full text]