We provide IT Staff Augmentation Services!

Hadoop Developer Resume

0/5 (Submit Your Rating)

Evanston, IL

SUMMARY

  • Over 16 years of Information Technology experience in Hadoop, HDFS, MapReduce programs, Informatica ETL design, analysis, Data Modeling, SQL, Relational and NoSQL Databases.
  • Excellent in Apache Hadoop (CDH3/CDH4) architecture and HDFS framework.
  • Experience in Map - Reduce programming model including Hive, Pig, HBase, Sqoop, Oozie, Flume
  • Experience in installing, configuring and using ecosystem components like Hadoop, MapReduce,
  • HDFS, Pig & Hive.
  • Good knowledge of Hadoop architecture and it’s components like HDFS, Job Tracker, Task Tracker,
  • Name Node, Data Node and Map-Reduce concepts.
  • Experience in Exporting/Importing data from using Sqoop from HDFS to RDBMS and vice-versa.
  • Worked in analyzing data using HIVEQL, PIG Latin and Map-Reduce programs in JAVA.
  • Good experience in Normalization, Denormalization and Multi Dimensional design Techniques.
  • Have significant expertise in all aspects of OLTP, OLAP processing.
  • Experience in PL/SQL, T-SQL, SQL, Unix Shell Scripting, Windows Shell Scripting, Perl, Awk, Korn
  • Have the quality to define user’s need, team player, and a great aptitude to learn quickly.
  • A self-motivated professional and natural communicator possessing good technical, initiating, leadership and problem-solving skills and has proven to be a good team player.

TECHNICAL SKILLS

Database: Apache Hbase, Oracle 11g/Exadata, Teradata, Bteq

ETL Tool: Informatica Power center 9.5

BI Tools: Business Objects XI, Crystal Reports XI

Other Tools: Erwin 8.x, TOAD 9.x, Visio, Ms-word, Excel

O/S: UNIX AIX, Sun Solaris, MVS, Windows 7, DOS

Language: MapReduce (0.20 YARN), HIVEQL, PIGLATIN, Core JAVA, JavaScript, SQL, PL/SQL, T-SQL, UNIX Shell Scripting.

Big Data: Apache Hadoop Architecture, HDFS, MapReduce Programming Model, HBase, Pig, Hive, Sqoop, ZooKeeper

Virtual Machine: VMWare, Virtual Box

PROFESSIONAL EXPERIENCE

Confidential

Hadoop Developer

Responsibilities:

  • Actively participated in software development lifecycle (scope, design, implement, deploy, test), including design and code reviews, test development, test automation.
  • Developed, tested, deployed, and operationally supported technical solutions in one or moreBI/DW areas -- such as Developing Map and Reduce jobs in Core Java, Creating Tables in HBaseExporting data from HBase to HDFS and to Oracle.
  • Developed Map-Reduce jobs to process the data and necessary HFiles.
  • Involved in performance enhancements of the code by writing custom comparators and combiner logic.
  • Involved in loading the HFiles into HBase for faster access of large customer base.
  • Documented the system processes and procedures for future .
  • Assisted in performing unit testing of Map-Reduce jobs using MRUnit.
  • Used Oozie scheduler to automate the pipeline workflow and orchestrate the map reduce jobs thatExtract the data on a timely manner.
  • Used ZooKeeper for providing coordinating services to the cluster.
  • Involved in story-driven agile development methodology and actively participated in daily scrummeetings.

Environment: Hadoop 2.x, HDFS, HBase, Oozie, Sqoop, Flume, Informatica 9.5, Oracle Exadata, CoreFTP,, Core Java, MRUnit, Eclipse, Bash Shell Scripting, ZooKeeper, Maven, MS office

Confidential

Hadoop Developer

Responsibilities:

  • Installed and configured Hadoop Map Reduce, HDFS and other Apache ecosystems
  • Analysis and understanding of business requirements.
  • Have written UNIX shell scripts to stream line the Data files loading for HDFS.
  • Involved in running Hadoop jobs for processing millions of records of text data
  • Developed MapReduce application using Hadoop, MapReduce programming and Hbase.
  • Loaded data in parallel processing.
  • Experience working in Agile development Team.

Environment: Hadoop HDFS, Map/Reduce, Informatica 8.6.1 (power center/power mart), Oracle 11g, WIN SCP, Sqoop, Sun Solaris Unix, Erwin 8, Java, JUNIT, Eclipse, MS office, MVN.

Confidential, Evanston, IL

Sr. Software Engineer

Responsibilities:

  • Analyzed and fixed P1 issues during the data load for CRMview tool on weekly basis.
  • Analyzed data and business rules to resolve P2 and P3 issues and incorporating solutions into new
  • Releases of CRMview tool as a fix.
  • Prepared design document for the new releases of CRMview tool.
  • Worked with BA to make sure all development aspects from PSD (proposed solution document) are covered in Design spec document.
  • Prepared Informatica mapping design specifications for the releases of crmView tool.
  • Created Visio diagrams to visualize process flow and data flow for the new development work.
  • Developed informatica mappings, mapplets, workflows, sessions.
  • Tested informatica jobs and prepared unit test documents before roll over to QA.
  • Entered Test cases in the Mercury Quality Center tool.
  • Prepared deployment notes for successful deployment from DEV to UAT & PROD.
  • Written PL/SQL procedures and functions to make use of them in the informatica mappings.
  • Written Shell scripts to call them in the command task session in the workflow manager.
  • Analyzed oracle table structures as part of the P2 and P3 issues to fix and suggesting changes.
  • Designed new informatica jobs and re-designed the existing informatica jobs to adjust to the ever
  • Changing business needs.
  • Designed informatica audit tables in the re-design of the informatica data loading for CRMView
  • Tool.
  • Generated DDL in Erwin tool and submitted to DBA for table creation.
  • Participating in production support whenever needed.

Environment: Informatica 8.6 (power center/power mart), Oracle 10g, Toad 9.x, WIN SCPCognos BI Reporting tool 8.3, CA Erwin 7.3, Sun Solaris, Linux, Share Point, Win CVS, MS office.

Confidential, Folsom, CA

Data Modeler

Responsibilities:

  • Understood Business Requirements from SME and understanding Ticket Types, Discount, Loyalty Points, and POS Transactions Subject Areas.
  • Understood the data of the Source applications Rentrack, Oracle DPA and AMCHO.
  • Prepared documentation with the business rules to transform data from Source to Target structures.
  • Involved in logical modeling of the MDM (Master Data Management) database and Radiant table structures for the EDW Pilot project.
  • Created Logical Model for Ticket Types, Discount, Loyalty Points, and Transactions Subject Areas.
  • Generated DDL in ER Studio and Created Physical Structures in database.
  • Created Indexes, Triggers, and Stored Procedures in database as per the requirements.
  • Prepared up to date documentation to keep informed the audience with the latest changes in the database.
  • Experience in integration of various data sources like Oracle, SQL server and flat files into staging area.
  • Explained logical models and business rules to the ETL Team.
  • Participated in the preparation of test cases in order to check data consistency, data quality

Environment: Embarcadero ER-Studio 7.5, SQL Server 2005, Oracle 10G, Toad 9.7SqlServer Management Studio 2005, Informatica 8.5, DataStage 8.x, SharePoint, MS-Word, Excel

Confidential

Sr Informatica Developer and Analyst

Responsibilities:

  • Participated in the Informatica Administration, Maintenance and Support.
  • Participated in development requirement meetings with SME to understand data loading needs.
  • Created Connections to access data from source tables in a relational and adabas databases.
  • Upgraded informatica 7.1 Mappings to informatica 8.1 as per the schedule.
  • Written Shell script and scheduled jobs to backup repository database on daily basis.
  • Written shell scripts for file management tasks.
  • Understanding Logical, Physical Modeling and entity relations from AdminServer application lead.
  • Understanding source data and file structures for loading data into target data storage structures.
  • Created Mappings using Source Qualifier, Aggregator, Expression, Joiner, Lookup
  • (connected/unconnected), Router, updated strategy, Normalizer, Sorter, Filter, Application
  • Source Qualifier, XML Source Qualifier, XML Generator, XML Parser Transformations for populating target table in efficient manner.
  • Optimized Transformations, Mappings and sessions using Informatica optimization techniques.
  • Mapping and session level parameters, Variables are used in the ETL Development.
  • Did performance tuning of targets, sources, mappings and sessions using various components like parameter file, variables, and dynamic cache, round robin, hash, auto key, Key range and Database partitioning.
  • Wrote Teradata Macros and used various Teradata analytic functions.
  • Wrote, tested and implemented Teradata Fastload, Multiload and Bteq scripts, DML and DDL.
  • Performance tuned and optimized various complex SQL queries.
  • Participated in the Informatica Migration Requests from DEV to UAT and UAT to PROD.
  • Created, Updated and maintained ETL Metadata using Metadata Exchange.
  • Provided 24X7 production supports on a rotation basis.

Environment: Informatica 7.1.3/8.1 (Power Center/Power Mart), Shell Scripting, AIX UNIX 5.xIBM DB2, WinSQL, Teradata V2R6, BTEQ, Zephyr Mainframe GUI, TSO/ISPF, ADABAS, XML,XSL,XSLT, ESP Scheduler, Java, Erwin 7.x, Ms-Word, MS-Excel, Visio

Confidential, Newark, DE

Software Developer

Responsibilities:

  • External application data is extracted, transformed and loaded in EDW.
  • Java applications are built for data transformation purposes.

Environment: Oracle 8i, DB2UDB7.1, PL/SQL, SQL*Plus, Erwin3.6, UNIX SHELL Scripting, Java 2.0, JavaScript, CSS, XML

Confidential

Software Developer

Responsibilities:

  • Developed Client/Server applications and Web applications using JSP, Java 1.1, DHTML, HTML, CSS, JavaScript for Fintech System’s Clients in India.

Environment: Windows NT, Oracle8.0, PL/SQL, Java, JSP, HTML, JavaScript, HTML, AIX UNIX, Bourne and Korn Shell Scripting, UNIX Commands.

We'd love your feedback!