Indeed Resume

Total Page:16

File Type:pdf, Size:1020Kb

Load more

Aditya Dhavala Hadoop Developer - Sears Holding - Email me on Indeed: indeed.com/r/Aditya-Dhavala/4a1ec64204eb4bd3 • Around 7 years of IT experience as a Developer, Designer & quality reviewer with cross platform integration experience using Hadoop, Java, J2EE and SOA. • Hands on experience in installing, configuring and using Apache Hadoop ecosystems such as Map Reduce, HIVE, PIG, SQOOP, FLUME and OOZIE. • Hands on experience on Hortonworks and Cloudera Hadoop environments. • Strong understanding of Hadoop daemons and Map-Reduce concepts. • Experienced in importing-exporting data into HDFS format. • Experienced in analyzing big data using Hadoop environment. • Experienced in handling Hadoop Ecosystem Projects such as Hive, Pig and Sqoop. • Experienced in developing UDFs for Hive using Java. • Strong understanding of NoSQL databases like HBase, MongoDB & Cassandra. • Familiar with handling complex data processing jobs using Cascading. • Hands on experience with Hadoop, HDFS, MapReduce and Hadoop Ecosystem (Pig, Hive, Oozie, Flume and Hbase). • Extensive experience in design, development and support Model View Controller using Struts and Spring framework. • Develop reusable solution to maintain proper coding standard across different java project. • Proficiency with the application servers like WebSphere, WebLogic, JBOSS and Tomcat • Developed core modules in large cross-platform applications using JAVA, J2EE, Spring, Struts, Hibernate, JAX-WS Web Services, and JMS. • Expertise in debugging and optimizing Oracle and java performance tuning with strong knowledge in Oracle 11g and SQL • Ability to work effectively in cross-functional team environments and experience of providing training to business users. WORK EXPERIENCE Hadoop Developer Sears Holding - Hoffman Estates, IL - February 2013 to Present Sears Holdings Corporation is a leading integrated retailer with almost 2,500 full-line and specialty retail stores in the United States and Canada. Sears Holdings is the leading home appliance retailer as well as a leader in tools, lawn and garden, fitness equipment and automotive repair and maintenance. I was a part of Big Data Processing Team to take advantage of available user data to make better decisions that significantly enhanced organizational success. I was involvedin setting up Cloudera Hadoop Cluster and wrote MapReduce jobs, Hive queriesand PigLatin scripts to explore through the data of customer sales to find significant information for trend analysis. Responsibilities: • Involved in start to end process of hadoop cluster installation, configuration and monitoring. • Responsible for building scalable distributed data solutions using Hadoop • Installed and configured Hive, Pig, Sqoop, Flume and Oozie on the Hadoop cluster • Setup and benchmarked Hadoop/HBase clusters for internal use • Developed Simple to complex Map/reduce Jobs using Hive and Pig • Optimized Map/Reduce Jobs to use HDFS efficiently by using various compression mechanisms • Handled importing of data from various data sources, performed transformations using Hive, MapReduce, loaded data into HDFS and Extracted the data from MySQL into HDFS using Sqoop • Analyzed the data by performing Hive queries and running Pig scripts to study customer behavior • Used UDF's to implement business logic in Hadoop • Implemented business logic by writing UDFs in Java and used various UDFs from Piggybanks and other sources. • Continuous monitoring and managing the Hadoop cluster using Cloudera Manager • Worked with application teams to install operating system, Hadoop updates, patches, version upgrades as required • Installed Oozieworkflow engine to run multiple Hive and Pig jobs • Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports for the BI team Environment: Hadoop, MapReduce, HDFS, Hive, Pig, Java, SQL, Cloudera Manager, Sqoop, Flume, Oozie, Java (jdk 1.6), Eclipse Hadoop Developer American Express - Phoenix, AZ - February 2012 to January 2013 American Express provides innovative payment, travel and expense management solutions for individuals and businesses of all sizes. It helps customers realize their dreams and aspirations through industry-leading benefits, access to unique experiences, business-building insights, and global customer care. Purpose of the project is to create Enterprise Data Hub so that various business units and use the date from Hadoop to do Data Analytics.The solution is based on the Cloudera Hadoop. The data will be stored in Hadoop file system and processed using Map/Reduce jobs. Responsibilities: • Installed and configured Hadoop MapReduce, HDFS and developed multiple MapReduce jobs in Java for data cleansing and preprocessing. • Involved in loading data from UNIX file system to HDFS. • Installed and configured Hive and also written Hive UDFs. • Evaluated business requirements and prepared detailed specifications that follow project guidelines required to develop written programs. • Devised procedures that solve complex business problems with due considerations for hardware/software capacity and limitations, operating times and desired results. • Analyzed large amounts of data sets to determine optimal way to aggregate and report on it. • Provided quick response to ad hoc internal and external client requests for data and experienced in creating ad hoc reports. • Responsible for building scalable distributed data solutions using Hadoop. • Responsible for cluster maintenance, adding and removing cluster nodes, cluster monitoring and troubleshooting, manage and review data backups, manage and review Hadoop log files. • Worked hands on with ETL process. • Handled importing of data from various data sources, performed transformations using Hive, MapReduce, and loaded data into HDFS. • Extracted the data from Teradata into HDFS using Sqoop. • Analyzed the data by performing Hive queries and running Pig scripts to know user behavior like shopping enthusiasts, travelers, music lovers etc. • Exported the patterns analyzed back into Teradata using Sqoop. • Continuous monitoring and managing the Hadoop cluster through Cloudera Manager. • Installed Oozie workflow engine to run multiple Hive. • Developed Hive queries to process the data and generate the data cubes for visualizing. Environment: Hadoop, MapReduce, HDFS, Hive, Ooozie, Java (jdk1.6), Cloudera, NoSQL, Oracle 11g, 10g, PL SQL, SQL*PLUS, Toad 9.6, Windows NT, UNIX Shell Scripting. Java/J2EE Application Developer Columbia Bank - Lakewood, WA - August 2010 to January 2012 The Columbia Bank Card Verification system is assessment system for credit card application. It processes the application of the customer who wants a credit card and processes till it gets either accepted or rejected. Responsibilities: • Involved in Java, J2EE, struts, web services and Hibernate in a fast paced development environment. • Followed agile methodology, interacted directly with the client provide/take feedback on the features, suggest/ implement optimal solutions, and tailor application to customer needs. • Rich experiences of database design and hands on experience of large database systems: Oracle 8i and Oracle 9i. • Involved in design and implementation of web tier using Servlets and JSP. • Used Apache POI for Excel files reading. • Written build scripts with Ant for deploying war and ear applications. • Developed the user interface using JSP and Java Script to view all online trading transactions. • Designed and developed Data Access Objects (DAO) to access the database. • Used DAO Factory and value object design patterns to organize and integrate the JAVA Objects • Coded Java Server Pages for the Dynamic front end content that use Servlets and EJBs. • Coded HTML pages using CSS for static content generation with JavaScript for validations. • Used JDBC API to connect to the database and carry out database operations. • Used JSP and JSTL Tag Libraries for developing User Interface components. • Performing Code Reviews. • Performed unit testing, system testing and integration testing. • Involved in building and deployment of application in Linux environment. • Deploying application in Development and Production servers. Environment: Java, J2EE, JDBC, Struts, SQL language. Hibernate, Eclipse, Apache POI, CSS. Java/J2EE Developer Capital One Bank - McLean, VA - October 2009 to June 2010 CapitalOne Auto Finance (COAF) is project where we design an application which can be used by the CapitalOne Bank to deal with different types of Auto loans depending upon the Customers Eligibility. COAF includes the maintaining of the existing Application. Responsibilities: • Played the role of Java developer in the project called "Coverage Selection Tool". • Technologies involved are EJB 3.0, Web services, Dojo (UI Framework) and other J2EE server components. • Analyze and prepare technical specifications with UML diagrams (Use case, Class, and Sequence diagrams • Used Rational Rose to develop the components required by client. • Wrote complex logic for forecasting the price of the products and subparts in next future quarters. • Development of business components applying OOAD and using good design patterns like, DAO, Value Objects, DTO, Factory, singleton. • Implemented DOM parsing module and created XSD and XSLT components. • Used stored procedures and Triggers extensively to develop the Backend business logic in Oracle database. • Involved in performance improving and bug fixing. • Analyze old database table fields and map to new schema tables using complex SQL Queries
Recommended publications
  • Commonjavajars - a Package with Useful Libraries for Java Guis

    Commonjavajars - a Package with Useful Libraries for Java Guis

    CommonJavaJars - A package with useful libraries for Java GUIs To reduce the package size of other R packages with Java GUIs and to reduce jar file conflicts, this package provides a few commonly used Java libraries. You should be able to load them by calling the rJava .jpackage function (a good place is most likely the .onLoad function of your package): .jpackage("CommonJavaJars", jars=c("forms-1.2.0.jar", "iText-2.1.4.jar")) We provide the following Java libraries: Apache Commons Logging under the Apache License, Version 2.0, January 2004, http://commons. apache.org/logging/, Copyright 2001-2007 The Apache Software Foundation Apache jog4j under Apache License 2.0, http://logging.apache.org/log4j/, Copyright 2007 The Apache Software Foundation Apache Commons Lang under Apache License 2.0, http://commons.apache.org/lang/, Copyright 2001-2011 The Apache Software Foundation Apache POI under Apache License 2.0, http://poi.apache.org/, Copyright 2001-2007 The Apache Software Foundation Apache Commons Collections under the Apache License, Version 2.0, January 2004, http://commons. apache.org/collections/, Copyright 2001-2008 The Apache Software Foundation Apache Commons Validator under the Apache License, Version 2.0, January 2004, http://commons. apache.org/validator/, Copyright 2001-2010 The Apache Software Foundation JLaTeXMath under GPL >= 2.0, http://forge.scilab.org/index.php/p/jlatexmath/, Copyright 2004-2007, 2009 Calixte, Coolsaet, Cleemput, Vermeulen and Universiteit Gent iText 2.1.4 under LGPL, http://itextpdf.com/, Copyright
  • Merchandise Planning and Optimization Licensing Information

    Merchandise Planning and Optimization Licensing Information

    Oracle® Retail Merchandise Planning and Optimization Licensing Information July 2009 This document provides licensing information for all the third-party applications used by the following Oracle Retail applications: ■ Oracle Retail Clearance Optimization Engine ■ Oracle Retail Markdown Optimization ■ Oracle Retail Place ■ Oracle Retail Plan ■ Oracle Retail Promote (PPO and PI) Prerequisite Softwares and Licenses Oracle Retail products depend on the installation of certain essential products (with commercial licenses), but the company does not bundle these third-party products within its own installation media. Acquisition of licenses for these products should be handled directly with the vendor. The following products are not distributed along with the Oracle Retail product installation media: ® ■ BEA WebLogic Server (http://www.bea.com) ™ ■ MicroStrategy Desktop (http://www.microstrategy.com) ■ MicroStrategy Intelligence Server™ and Web Universal (http://www.microstrategy.com) ® ■ Oracle Database 10g (http://www.oracle.com) ® ■ Oracle Application Server 10g (http://www.oracle.com) ® ■ Oracle Business Intelligence Suite Enterprise Edition Version 10 (http://www.oracle.com) ■ rsync (http://samba.anu.edu.au/rsync/). See rsync License. 1 Softwares and Licenses Bundled with Oracle Retail Products The following third party products are bundled along with the Oracle Retail product code and Oracle has acquired the necessary licenses to bundle the software along with the Oracle Retail product: ■ addObject.com NLSTree Professional version 2.3
  • Combined Documents V2

    Combined Documents V2

    Outline: Combining Brainstorming Deliverables Table of Contents 1. Introduction and Definition 2. Reference Architecture and Taxonomy 3. Requirements, Gap Analysis, and Suggested Best Practices 4. Future Directions and Roadmap 5. Security and Privacy - 10 Top Challenges 6. Conclusions and General Advice Appendix A. Terminology Glossary Appendix B. Solutions Glossary Appendix C. Use Case Examples Appendix D. Actors and Roles 1. Introduction and Definition The purpose of this outline is to illustrate how some initial brainstorming documents might be pulled together into an integrated deliverable. The outline will follow the diagram below. Section 1 introduces a definition of Big Data. An extended terminology Glossary is found in Appendix A. In section 2, a Reference Architecture diagram is presented followed by a taxonomy describing and extending the elements of the Reference Architecture. Section 3 maps requirements from use case building blocks to the Reference Architecture. A description of the requirement, a gap analysis, and suggested best practice is included with each mapping. In Section 4 future improvements in Big Data technology are mapped to the Reference Architecture. An initial Technology Roadmap is created on the requirements and gap analysis in Section 3 and the expected future improvements from Section 4. Section 5 is a placeholder for an extended discussion of Security and Privacy. Section 6 gives an example of some general advice. The Appendices provide Big Data terminology and solutions glossaries, Use Case Examples, and some possible Actors and Roles. Big Data Definition - “Big Data refers to the new technologies and applications introduced to handle increasing Volumes of data while enhancing data utilization capabilities such as Variety, Velocity, Variability, Veracity, and Value.” The key attribute is the large Volume of data available that forces horizontal scalability of storage and processing and has implications for all the other V-attributes.
  • Pharmacy Product System – National (Pps-N) Installation Guide

    Pharmacy Product System – National (Pps-N) Installation Guide

    PHARMACY PRODUCT SYSTEM – NATIONAL (PPS-N) INSTALLATION GUIDE December 2016 Version 1.2 Department of Veterans Affairs Office of Information and Technology (OIT) PPS-N Installation Guide v1.2 i December 2016 Revision History Date Version Revised Description Author Pages November 1.2 All Updated content with installation REDACTED. HPE 2016 instructions for Fixed Medication Copay FMCT Team. Tiers (FMCT) Release 1.2. May 2015 1.1.02 Updated date and version number to 1.1.02. Enterprise Updated the PPS-N EAR file name. Application Maintenance August 1.1.01 Updated version number to 1.1.01, updated Enterprise 2014 the PPS-N EAR file name and the PPSNS Application MUMPS KIDS file name. Maintenance Added instructions to Undeploy the application. And made some formatting changes. November 1.0.01 Updated version number to 1.0.01, updated Enterprise 2013 the PPS-N EAR file name and the PPSNS Application MUMPS KIDS file name. Maintenance January 1.0 Updated document to modify formatting SwRI 2013 based on NRR Review. December 1.0 No applicable updates for this document SwRI 2012 November 1.0 Updated section 10.5.1 to include a SwRI 2012 reference to other applications updating the image folder October 1.0 Version 1.0 updates SwRI 2012 September 1.0 Version 1.0 SwRI 2012 PPS-N Installation Guide v1.2 ii December 2016 TABLE OF CONTENTS 1 PROJECT SCOPE ....................................................................................................................... 1 1.1 Project Identification .....................................................................................................................
  • Oracle® Application Management Pack for Oracle E-Business Suite Guide Release 12.1.0.2.0 Part No

    Oracle® Application Management Pack for Oracle E-Business Suite Guide Release 12.1.0.2.0 Part No

    Oracle® Application Management Pack for Oracle E-Business Suite Guide Release 12.1.0.2.0 Part No. E39873-01 November 2013 Oracle Application Management Pack for Oracle E-Business Suite Guide, Release 12.1.0.2.0 Part No. E39873-01 Copyright © 2007, 2013, Oracle and/or its affiliates. All rights reserved. Primary Author: Biju Mohan, Mildred Wang Contributing Author: Srikrishna Bandi, KrishnaKumar Nair, Angelo Rosado Contributor: John Aedo, Max Arderius, Kenneth Baxter, Bhargava Chinthoju, Lauren Cohn, Rumeli Das, Ivo Dujmovic, Nipun Goel, Clara Jaeckel, Ryan Landowski, Senthil Madhappan, Biplab Nayak, Shravankumar Nethi, Vinitha Rajan, Vasu Rao, Traci Short, Mike Smith Oracle and Java are registered trademarks of Oracle and/or its affiliates. Other names may be trademarks of their respective owners. Intel and Intel Xeon are trademarks or registered trademarks of Intel Corporation. All SPARC trademarks are used under license and are trademarks or registered trademarks of SPARC International, Inc. AMD, Opteron, the AMD logo, and the AMD Opteron logo are trademarks or registered trademarks of Advanced Micro Devices. UNIX is a registered trademark of The Open Group. This software and related documentation are provided under a license agreement containing restrictions on use and disclosure and are protected by intellectual property laws. Except as expressly permitted in your license agreement or allowed by law, you may not use, copy, reproduce, translate, broadcast, modify, license, transmit, distribute, exhibit, perform, publish, or display any part, in any form, or by any means. Reverse engineering, disassembly, or decompilation of this software, unless required by law for interoperability, is prohibited. The information contained herein is subject to change without notice and is not warranted to be error-free.
  • Apache Sentry

    Apache Sentry

    Apache Sentry Prasad Mujumdar [email protected] [email protected] Agenda ● Various aspects of data security ● Apache Sentry for authorization ● Key concepts of Apache Sentry ● Sentry features ● Sentry architecture ● Integration with Hadoop ecosystem ● Sentry administration ● Future plans ● Demo ● Questions Who am I • Software engineer at Cloudera • Committer and PPMC member of Apache Sentry • also for Apache Hive and Apache Flume • Part of the the original team that started Sentry work Aspects of security Perimeter Access Visibility Data Authentication Authorization Audit, Lineage Encryption, what user can do data origin, usage Kerberos, LDAP/AD Masking with data Data access Access ● Provide user access to data Authorization ● Manage access policies what user can do ● Provide role based access with data Agenda ● Various aspects of data security ● Apache Sentry for authorization ● Key concepts of Apache Sentry ● Sentry features ● Sentry architecture ● Integration with Hadoop ecosystem ● Sentry administration ● Future plans ● Demo ● Questions Apache Sentry (Incubating) Unified Authorization module for Hadoop Unlocks Key RBAC Requirements Secure, fine-grained, role-based authorization Multi-tenant administration Enforce a common set of policies across multiple data access path in Hadoop. Key Capabilities of Sentry Fine-Grained Authorization Permissions on object hierarchie. Eg, Database, Table, Columns Role-Based Authorization Support for role templetes to manage authorization for a large set of users and data objects Multi Tanent Administration
  • What's New with Apache POI

    What's New with Apache POI

    What's new with Apache POI Apache POI - The Open Source Java solution for Microsoft Office Nick Burch Senior Developer Torchbox Ltd What'll we be looking at? • POI and OOXML – why bother? • Working with Excel files – DOM-like UserModel for .xls and .xlsx – SAX-like EventModel for .xls • Converting older code to the new style • PowerPoint – ppt and pptx • Word, Visio, Outlook, Publisher But first... OLE2 & OOXML • All the old style file formats (xls, doc, ppt, vsd) are OLE2 based • OLE2 is a binary format, that looks a lot like FAT • All the new file formats (xlsx, docx, pptx) are OOXML • OOXML is a zip file of XML files, which are data and metadata OLE2 Overview • OLE 2 Compound Document Format • Binary file format, developed by Microsoft • Contains different streams / files • POIFS is our implementation of it • http://poi.apache.org/poifs/fileformat.html has the full details • OOXML structure nods back to OLE2 Peeking inside OLE2 • POIFS has several tools for viewing and debugging OLE2 files • org.apache.poi.poifs.dev. POIFSViewer is the main one • Ant task is “POIFSViewer”, pass it -Dfilename=src/testcase/.... • Documents have metadata entries, and streams for their data SimpleWithColour.xls POIFS FileSystem Property: "Root Entry" Name = "Root Entry" DocumentSummaryInformation Property: "DocumentSummaryInformation" Name = "DocumentSummaryInformation" Document: "DocumentSummaryInformation" size=261 SummaryInformation Property: "SummaryInformation" Name = "SummaryInformation" Document: "SummaryInformation" size=229 Workbook Property:
  • Apache Ant Best Practices

    Apache Ant Best Practices

    08_Lee_ch05.qxd 5/3/06 5:12 PM Page 81 C HAPTER 5 Apache Ant Best Practices This chapter looks in more detail at some best practices for using Ant on real projects. First I describe the use of property files to enable configuration of the build process depending on a user’s role and requirements. I then describe how best to integrate Ant with IBM Rational ClearCase. Finally, I look at some general best practices for supporting the build process on large projects. Aims of This Chapter Apache Ant is a powerful build tool with significant built-in capabilities. However, a few capabil- ities and best practices stand out; they are described here. After reading this chapter, you will be able to • Understand what Ant property files are and how they can be used to make build scripts more maintainable. • Understand how to use Ant’s capabilities to better integrate with IBM Rational ClearCase. • Implement Ant build files that support reuse and maintainability on large projects. This chapter assumes that you are familiar with the basic concepts of Apache Ant that were discussed in Chapter 4, “Defining Your Build and Release Scripts.” Property Files From the perspective of Chapter 4, an Ant build.xml file is a single centralized build file that defines a repeatable process for bringing together an application, usually producing some form of 81 08_Lee_ch05.qxd 5/3/06 5:12 PM Page 82 82 Chapter 5 Apache Ant Best Practices executable output. Although a single build.xml file can be enough to drive the build process, in practice it can quickly become large and unwieldy.
  • Talend Open Studio for Big Data Release Notes

    Talend Open Studio for Big Data Release Notes

    Talend Open Studio for Big Data Release Notes 6.0.0 Talend Open Studio for Big Data Adapted for v6.0.0. Supersedes previous releases. Publication date July 2, 2015 Copyleft This documentation is provided under the terms of the Creative Commons Public License (CCPL). For more information about what you can and cannot do with this documentation in accordance with the CCPL, please read: http://creativecommons.org/licenses/by-nc-sa/2.0/ Notices Talend is a trademark of Talend, Inc. All brands, product names, company names, trademarks and service marks are the properties of their respective owners. License Agreement The software described in this documentation is licensed under the Apache License, Version 2.0 (the "License"); you may not use this software except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0.html. Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. This product includes software developed at AOP Alliance (Java/J2EE AOP standards), ASM, Amazon, AntlR, Apache ActiveMQ, Apache Ant, Apache Avro, Apache Axiom, Apache Axis, Apache Axis 2, Apache Batik, Apache CXF, Apache Cassandra, Apache Chemistry, Apache Common Http Client, Apache Common Http Core, Apache Commons, Apache Commons Bcel, Apache Commons JxPath, Apache
  • Apache Open Office Spreadsheet Templates

    Apache Open Office Spreadsheet Templates

    Apache Open Office Spreadsheet Templates Practicing and publishable Lev still reasserts his administrator pithily. Spindle-legged Lancelot robotize or mention some thingumbob Bradypastorally, weekends however imminently. defenseless Dru beheld headforemost or lipped. Tempest-tossed Morris lapidifies some extravasation after glamorous Get familiar with complete the following framework which to publish a spreadsheet templates can even free and capable of the language id is this website extensions Draw is anchor on three same plague as Adobe Illustrator or Photoshop, but turning an announcement to anywhere to friends and grease with smart software still be ideal. Get started in minutes to try Asana. So much the contents of their own voting power or edit them out how do it is where can! Retouch skin problems. But is make it is done in writer blogs or round off he has collaborative effort while presenting their processes to learn how. Work environment different languages a lot? Layout view combines a desktop publishing environment so familiar Word features, giving have a customized workspace designed to simplify complex layouts. Enjoy finger painting with numerous colors that care can choose. Green invoice template opens a office, spreadsheets to the. Google docs and open office. Each office templates to open in a template opens in the darkest locations in critical situations regarding medical letter of. You open office templates are there is a template to apache open office on spreadsheets, and interactive tool with. Its print are produced a banner selling ms word document author in order to alternatives that. Manage Office programs templates Office Microsoft Docs. It includes just let every name you mean ever ask soon as a writer or editor.
  • Realization of Big Data Ana- Lytics Tool for Optimization Processes Within the Finnish Engineering Company

    Realization of Big Data Ana- Lytics Tool for Optimization Processes Within the Finnish Engineering Company

    OPINNÄYTETYÖ - AMMATTIKORKEAKOULUTUTKINTO TEKNIIKAN JA LIIKENTEEN ALA REALIZATION OF BIG DATA ANA- LYTICS TOOL FOR OPTIMIZATION PROCESSES WITHIN THE FINNISH ENGINEERING COMPANY A u t h o r / s : Karapetyan Karina SAVONIA UNIVERSITY OF APPLIED SCIENCES THESIS Abstract Field of Study Technology, Communication and Transport Degree Programme Degree Programme in Information Technology Author(s) Karapetyan Karina Title of Thesis Realization of Big Data Analytics Tool for optimization processes within the Finnish engineering company Date 23.05.2016 Pages/Appendices 54 Supervisor(s) Mr. Arto Toppinen, Principal Lecturer at Savonia University of Applied Sciences, Mr. Anssi Suhonen, Lecturer at Savonia University of Applied Sciences Client Organisation /Partners Hydroline Oy Abstract Big Data Analytics Tool offers an entire business picture for making both operational and strategic deci- sions from selecting the product price to establishing the priorities for the further vendor’s enhancement. The purpose of the thesis was to explore the industrial system of Hydroline Oy and provide a software solution for the elaboration of the manufacture, due to the internal analyzing within the company. For the development of Big Data Analytics Tool, several software programs and tools were employed. Java-written server controls all components in the project and visualizes the processed data via a user- friendly client web application. The SQL Server maintains data, observed from the ERP system. Moreo- ver, it is responsible for the login and registration procedure to enforce the information security. In the Hadoop environment, two research methods were implemented. The Overall Equipment Effectiveness model investigated the production data to obtain daily, monthly and annual efficiency indices of equip- ment utilization, employees’ workload, resource management, quality degree, among others.
  • Mapreduce Service

    Mapreduce Service

    MapReduce Service Troubleshooting Issue 01 Date 2021-03-03 HUAWEI TECHNOLOGIES CO., LTD. Copyright © Huawei Technologies Co., Ltd. 2021. All rights reserved. No part of this document may be reproduced or transmitted in any form or by any means without prior written consent of Huawei Technologies Co., Ltd. Trademarks and Permissions and other Huawei trademarks are trademarks of Huawei Technologies Co., Ltd. All other trademarks and trade names mentioned in this document are the property of their respective holders. Notice The purchased products, services and features are stipulated by the contract made between Huawei and the customer. All or part of the products, services and features described in this document may not be within the purchase scope or the usage scope. Unless otherwise specified in the contract, all statements, information, and recommendations in this document are provided "AS IS" without warranties, guarantees or representations of any kind, either express or implied. The information in this document is subject to change without notice. Every effort has been made in the preparation of this document to ensure accuracy of the contents, but all statements, information, and recommendations in this document do not constitute a warranty of any kind, express or implied. Issue 01 (2021-03-03) Copyright © Huawei Technologies Co., Ltd. i MapReduce Service Troubleshooting Contents Contents 1 Account Passwords.................................................................................................................. 1 1.1 Resetting