Hadoop Developer Resume
FL
SUMMARY:
- Overall 14+ years of IT experience in a variety of industries.
- My technical experience includes 4 years as a Hadoop Developer, and over 7 years on Hyperion products, and 4 years of QA.
- Throughout my career, I have worked on different domains such as Oil and Gas, Finance, Manufacturing and Services industries.
- Four years of Hadoop ecosystem experience in ingestion, storage, querying, processing and analysis of big data. The summary of various tasks performed using Hadoop ecosystem includes:
- Hands on experience in developing and deploying enterprise based applications using major components in Hadoop ecosystem like Hadoop Map Reduce, YARN, Hive, Pig, HBase, Flume, Sqoop, Kafka, Oozie and ZooKeeper.
- Hand on experience on Cloudera distribution
- Rich experience working with Shell Scripting
- Expertise in writing Hadoop Jobs for analyzing data using Hive QL (Queries), Pig Latin (Data flow language), and custom MapReduce programs in Java
- Wrote Hive queries for data analysis to meet the requirements
- Created Hive tables to store data into HDFS and processed data using Hive QL
- Hands on experience in designing and creating Hive tables using shared meta - store with partitioning and bucketing
- Expert in working with Hive data warehouse tool-creating tables, data distribution by implementing partitioning and bucketing, writing and optimizing the HiveQL queries
- Good knowledge in creating Custom Serdes in Hive
- Developed Pig Latin scripts using operators such as LOAD, STORE, DUMP, FILTER, DISTINCT, FOREACH, GENERATE, GROUP, COGROUP, ORDER, LIMIT, UNION, SPLIT to extract data from data files to load into HDFS
- Hands on experience in creating custom UDFs functions including UDAFs & UDTFs in Hive & Pig
- Hands on experience in using Pig for analysis on large unstructured data sets
- Automated Sqoop, Hive and Pig scripts using work flow scheduler Oozie, Spark and Kafka
- Experience in working with MapReduce programs using Apache Hadoop for working with Big Data
- Hands on experience in dealing with Compression Codecs like Snappy, Gzip.
- Good understanding of Data Mining and Machine Learning techniques
- Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems and vice-versa
- Hands on experience in configuring and working with Flume to load the data from multiple sources directly into HDFS
- Extensive experience with SQL, PL/SQL and database concepts
- Used Hbase in accordance with PIG/Hive as and when required for real time low latency queries.
- Knowledge of job workflow scheduling and monitoring tools like Oozie (hive, pig) and Zookeeper (Hbase)
- Experience in developing solutions to analyze large data sets efficiently
- Ability to work in high-pressure environments delivering to and managing stakeholder expectations
- Ability to adapt to evolving technology, strong sense of responsibility and .
SKILLS SUMMARY:
Hadoop: HDFS, HBase, MapReduce, YARN, Hive, Pig, Sqoop, Flume, Oozie, Kafka, Zookeeper, Cloudera and Hortonworks distributions
Oracle Platform and Architecture: Oracle 11g, Data Relationship Management 11.1.2.2,11.1.2.3,11.1.2.4, Oracle Portal
Non-Oracle Databases: SQL Server 2008, 2014
Tools: SQL Developer,PL/SQL Developer, JDeveloper 10g,11g
Hardware & Operating Systems: Windows 2008,Windows 2000, Windows 2003, HP-UNIX
Programming Languages: Java,Visual Basic, VBScript, HTML,JavaScript, PL / SQL
Engagement Experience: Applications Design/Development, Applications Support, Testing, Data Conversion, End User, Offshore
WORK EXPERIENCE:
Hadoop Developer
Confidential, FL
Responsibilities:
- Developed data pipeline using Flume, Sqoop, Pig and Java map reduce to ingest customer behavioral data and financial histories into HDFS for analysis
- Involved in writing MapReduce jobs
- Involved in SQOOP, HDFS Put or CopyFromLocal to ingest data
- Used Pig to do transformations, event joins, filter bot traffic and some pre-aggregations before storing the data onto HDFS
- Involved in developing Pig UDFs for the needed functionality that is not out of the box available from Apache Pig
- Used Hive to analyze the partitioned and bucketed data and compute various metrics for reporting
- Involved in developing Hive DDLs to create, alter and drop Hive TABLES
- Involved in developing Hive UDFs for the needed functionality that is not out of the box available from Apache Hive
- Involved in using HCATALOG to access Hive table metadata from Map Reduce or Pig code
- Computed various metrics using Java MapReduce to calculate metrics that define user experience, revenue etc
- Responsible for developing data pipeline using flume, sqoop and pig to extract the data from weblogs and store in HDFS Designed and implemented various metrics that can statistically signify the success of the experiment
- Used Eclipse and ant to build the application
- Involved in using SQOOP for importing and exporting data into HDFS and Hive
- Involved in processing ingested raw data using MapReduce, Apache Pig and Hive
- Involved in developing Pig Scripts for change data capture and delta record processing between newly arrived data and already existing data in HDFS
- Involved in pivot the HDFS data from Rows to Columns and Columns to Rows
- Involved in emitting processed data from Hadoop to relational databases or external file systems using SQOOP, HDFS GET or CopyToLocal
- Involved in developing Shell scripts to orchestrate execution of all other scripts (Pig, Hive, MapReduce) and move the data files within and outside of HDFS
Environment: Cloudera HDFS, MapReduce, Yarn, Hive, Pig, HBase, Oozie, Sqoop, Flume, Oracle 11g, Core Java, Eclipse.
Hadoop Developer
Confidential, Quincy, MA
Responsibilities:
- Responsible for coding Map Reduce program, Hive queries, testing and debugging the Map Reduce programs
- Responsible for Installing, Configuring and Managing of Hadoop Cluster spanning multiple racks
- Developed Pig latin scripts in the areas where extensive coding needs to be reduced to analyze large data sets
- Used Sqoop tool to extract data from a relational database into Hadoop
- Involved in performance enhancements of the code and optimization by writing custom comparators and combiner logic
- Worked closely with data warehouse architect and business intelligence analyst to develop solutions
- Good understanding of job schedulers like Fair Scheduler which assigns resources to jobs such that all jobs get, on average, an equal share of resources over time and an idea about Capacity Scheduler
- Responsible for performing peer code reviews, troubleshooting issues and maintaining status report
- Involved in creating Hive Tables, loading with data and writing Hive queries, which will invoke and run MapReduce jobs in the backend
- Involved in identifying possible ways to improve the efficiency of the system. Involved in the requirement analysis, design, development and Unit Testing use of MRUnit and Junit
- Prepare daily and weekly project status report and share it with the client
- Supported in setting up QA environment and updating configurations for implementing scripts with Pig, Hive and Sqoop
Environment: Apache Hadoop, Java (JDK 1.7), Oracle, My SQL, Cloudera HDFS, Hive, Pig, Sqoop, Linux, Cent OS, Junit, MRUnit
Confidential
Hadoop Developer / Administrator
Responsibilities:
- Installed/Configured/Maintained Apache Hadoop clusters for application development and Hadoop tools like Hive, Pig, HBase, Zookeeper and Sqoop.
- Installed cluster, commissioning & decommissioning of DataNodes, NameNode recovery, capacity and configured Hadoop, MapReduce, HDFS, developed multiple MapReduce jobs in JAVA for data cleaning.
- Wrote the shell scripts to monitor the health check of Hadoop daemon services and respond accordingly to any warning or failure conditions.
- Developed data pipeline using Flume, Sqoop, Pig and Java MapReduce to ingest customer behavioral data and financial histories into HDFS for analysis.
- Involved in collecting and aggregating large amounts of log data using Apache Flume and staging data in HDFS for further analysis.
- Collected the logs data from web servers and integrated in to HDFS using Flume.
- Worked on installing planning, and slots configuration.
- Implemented NameNode backup using NFS. This was done for High availability.
- Developed PIG Latin scripts to extract the data from the web server output files to load into HDFS.
- Installed Oozie workflow engine to run multiple Hive and Pig Jobs.
- Use of Sqoop to import and export data from HDFS to RDBMS and vice-versa.
- Used Hive and created Hive tables and involved in data loading and writing Hive UDFs.
- Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports.
- Involved in migration of ETL processes from Oracle to Hive to test the easy data manipulation.
- Worked on NoSQL databases including HBase, MongoDB, and Cassandra.
- Supported in setting up QA environment and updating configurations for implementing scripts with Pig and Sqoop.
Environment: NoSQL, Cassandra, MongoDB, Hortonworks HDFS, HBase, PIG Latin, Hive, Flume, MapReduce, Sqoop, JAVA, Eclipse, NetBeans.
Confidential, Houston, TX
Sr. Hyperion Infrastructure Administrator-Hyperion
Responsibilities:
- Administration of Production, QA, Archive& Development landscape which includes installation, maintenance and performance tuning and user administration using the shared services.
- Provide System Admin support for HFM, HTP, FDM, Financial reports, Hyperion Essbase & Planning including support MEC processing, application testing and process documentation
- Managing Windows Servers (Monitoring resources on Windows Server, Starting and Stopping Hyperion services, Monitor app back-ups, manage resources ( CPU, disk space and RAM) and work with Windows server team adding additional resources whenever needed.
- Worked on the installations of EPM 11.1.2 versions and migrated applications from 9.x to 11.1.2.1, 11.1.2.3
- Production support which includes EPM (11.1.2.1) suite Hyperion Financial Management (HFM), HTP, Essbase and Hyperion Planning applications.
- Installation of EPM11.1.2, System 9 Shared services/License server, EAL, AAS, HFM 11.1.2, 9.3.1, 9.2.03& 4.1.1.1, HTP, planning 11.1.2, 9.2.03. & 4.1.1.1& BI reports 11.1.2, 9.3.1, 9.2.03. & 901. DRM 11.1.2& Calc Mgr
- Migration from System 9.3.1 version to EPM 11.1.2.1, 11.1.2.3( HFM, EAS,FDM, EAL& EPMA )
- Create/Maintain system administration scripts, and maintain responsibility for creating and updating system documentation.
- Migrating artifacts across the instances using LCM utility.
- Manage projects in Oracle Data Integrator to develop interfaces, manage repositories and other artifacts.
- Identify, implement and test patches and upgrades and maintain up-to-date documentation of Hyperion-related configuration.
- User administration and setting up MSAD, Groups/User security for Planning, HFM & BI reports in Shared services.
- Also involved in performance tuning for BI reports, HFM and Planning servers.
- Monitor application performance and availability. And worked closely with management and end-users to develop and enhance the application processes that will result in a stable productive environment.
- Facilitate the promotion of Hyperion objects to the Development and Production environments and create/Maintain system administration scripts, and maintain responsibility for creating and updating system documentation.
- Identify, implement and test patches and upgrades and maintain up-to-date documentation of Hyperion-related configuration.
- Fine tuning of BI+ workspace reports, HFM and Essbase which includes OS, application & web services tuning.
- Weekly server maintenance of Production, Test & Development servers which includes backup and recycling the services. Automation of Essbase backups using ESSCMD& Maxl scripts.
- Involved DR testing and setup.
- Setup the CITRIX environment for Planning &HFM for windows 2007 users.
- Troubleshoot issues with both client and server systems i.e. Smart View, Excel Add-in, Web Analysis, Essbase, Financial Reporting, Workspace, Studio, Web Logic
- Responsible for providing weekly/monthly updated information to IT internal security audit for SOX.
Environment: EPM11.1.2 (EAL, EAL, Workspace& FDM), System 9 shared services 11.1.2, 9.3.1, 9.2.0.3& 9.01, HFM 11.1.2, 9.3.1, 9.2.0.3& 4.1.1, HTP (Hyperion Tax Provisioning Administrator), Planning4.1.1 &9.2.0.3, EAS11.1.2, 9.3.1&9.01.1, Upstream, DRM11.1.2, ODI, FDM (9.3.1& 11.1.2). Oracle 10g&11g, Web logic 8.x & 9.x, Apache Tomcat, IIS 6, OEL (Linux5.7) and Windows-2003& 2008R2.
Confidential
QA Analyst
Responsibilities:
- Involved in the development of Test plan, Test case preparation, Test strategy and the functional analysis of the system
- Implemented HP Quality Center for Test Planning, Test Case writing, Test Execution
- Involved in defect tracking and reporting using Quality Center
- Document functional and non-functional technical requirements
- Performed Database Validation to check the updated data in the database using SQL queries using TOAD
- Involved in developing, executing and maintaining test scripts using QTP
- Generated detailed reports that include graphs and tables for various test results
- Evaluated and Reported Test Results & the overall progress periodically to the Project Management
- Provided assistance to Project managers to develop and maintaintesting schedules
- Worked closely with Release Management team for all the upcoming builds and releases
Environment: HP Quality Center, Qucik Test Professional, Windows, Java, SQL Server, HTML, MS Office, XML, Oracle and UNIX