We provide IT Staff Augmentation Services!

Hadoop Developer Resume

3.00/5 (Submit Your Rating)

Bellevue, WA

PROFILE SUMMARY:

  • 15+ years of experience in Project Management, Software Development & System Analysis with expertise in working with very large databases & data warehouses using OLTP systems, Data warehousing & BI tools
  • Hands on experience in Big Data Technologies (Hadoop Framework) and its eco systems like HDFS, MapReduce, HBase, HIVE, Pig Latin, Sqoop, Zookeeper, Spark & Oozie.
  • Expert in working on Oracle database, advanced concepts & Performance Tuning and Teradata & Informatica.
  • Expertise in concepts of end - to-end project planning and implementation from scope management in various environments viz. release based maintenance, custom application development, enterprise wide application deployment, testing support and quality management in adherence to international guidelines and norms
  • Hands on experience in Java, Oracle, SQL, PL/SQL, Developer 2000 Forms, Reports & Graphs, PL/SQL web cartridge, SQL * Loader tool using various formats and in large volumes and in UNIX and shell scripting
  • Implemented new Intelligent Data CUBE (Micro Strategy) using Hadoop and its eco system components
  • Implemented new warehouse (hive) using open source solution (i.e. Hadoop framework and its eco system components)
  • Having good knowledge in No SQL DB’s like Cassandra.
  • Having good functional knowledge in Online Affiliates System & International Export Accounting.
  • Demonstrated excellence in migrating nearly 65 Million player accounts along with various key factors to the Target Platform.
  • This includes various territorial regulations.
  • Evaluated the functional gaps between Confidential and Confidential when both the organizations are merged.
  • Implemented the Intelligent Data CUBE (Micro Strategy) for Marketing and Finance teams for their Analysis purpose.
  • Complete De-Normalized data of online affiliate system to the minute granular level.
  • Implemented the player fraud control, player Duplication, player Qualification, player Loyalty and Player Bonus System
  • Implemented Incremental data load using Informatica Power Center.
  • Involved in the Oracle performance tuning.
  • Conducted source data analysis - SQL server, Oracle, My SQL, IBM Informix and Flat Files.

TECHNICAL SKILLS:

Operating Systems: Windows, UNIX, Linux

Hadoop Eco Systems: HDFS, Map Reduce (M/R), Hive, HBase, Pig, Sqoop, Zookeeper, SOLR, Spark & Oozie

Hadoop Distributions: Cloudera 5.4.3, HDP 2.2.4

Data Base: Oracle, Teradata

ETL Tools: Informatica

Reporting Tools: Micro Strategy (MSTR)

Scripting Tools: UNIX Shell Scripting, Python

Programming Languages: Java, J2EE, XML

Designing Tools: ERWIN

GUI Tools: Developer 2000 Suite & PL/SQL Web Cartridge

Web Technologies: JSP, Servlets & JavaScript

Project Management Tools: Microsoft Project, PPM, RT& iPlan

Source Control Tools: SVN, CVS&VSS

DB tools: TOAD, PL/SQL Developer & SQL Developer

MANAGERIAL SKILLS:

  • Managing requirement gathering, system analysis and finalization of technical / functional specifications
  • Planning activities as scoping, estimation, tracking, risk management, delivery management& post implementation support
  • Implementing project plans within preset budgets and deadlines
  • Interacting with the senior business and the IT client contacts to provide team leadership & technical and functional expertise
  • Mapping client's requirements and providing them the best solutions involving evaluation of options, definition of scope of project, and finalization of project requirements
  • Working with various levels of management and users to help define information strategy and ensure its consistency with the overall business strategy
  • Executing end to end management of projects from conceptualization and visualization to technology mapping and final execution of projects
  • Implementing process to capture functional design and assisting business team to take educated decisions by explaining technical features
  • Developing architecture and technical work plan to ensure successful development & implementation of applications

PROFESSIONAL EXPERIENCE:

Hadoop Developer

Confidential, Bellevue, WA

Environment: HDP 2.2.4, MapReduce, HDFS, Pig, Hive, Oozie, Python, Spark and Java (jdk 1.6)

Responsibilities:

  • Implementing the NRT data ingestion using Spark and it’s RDD’s.
  • Handled JSON files for data ingestion.
  • Implementing the business logic in PIG
  • Exporting the data into reporting server (Teradata) using Sqoop
  • Production deployment support and execution
  • Supporting the offshore team and QAT.

Confidential, Bellevue, WA

Environment: HDP 2.2.4, MapReduce, HDFS, Sqoop, Pig, Hive, Oozie, SOLR, HBase and Java (jdk 1.6)

Hadoop Developer

Responsibilities:

  • Implemented the business logic in preparation section using PIG
  • Implemented the Late Arriving Dimension logic
  • Implemented the SCD1 and SCD2 logic
  • Supporting the UAT with client
  • Data Loads in production

Confidential, Sanfrancisco, CA

Environment: CDH 5.4.3, AWS, M/R, HDFS, Sqoop, Pig, Hive, Oozie, Impala, HBase, Spark and Java (jdk 1.6)

Hadoop Developer

Responsibilities:

  • Implemented data ingestion part using Sqoop, M/R and Spark
  • Resolving the issues from incremental load
  • Duplicate record issues are resolved using the HBase
  • Processing the data using Pig Latin scripting
  • Scheduling the jobs using Oozie
  • Data validation using shell and python scripting.
  • Data sanity using by paperwork from each data source.
  • Reviewing the log files
  • Coordinating with client and offshore team.

Confidential, Bellevue, WA

Environment: Apache Hadoop, M/R, HDFS, Spark,HBase, Hive, Pig Latin, Oozie and Java (jdk 1.6)

Hadoop Developer

Responsibilities:

  • Worked on HDP to analyze data present on top of HDFS.
  • Worked extensively on Hive and Pig Latin.
  • Worked on large sets of structured, semi-structured and unstructured data. Few data sets using Spark.
  • Developed Pig Latin scripts to play with the data.
  • Involved in creating Hive tables, loading with data and writing hive queries which will run internally in map reduce way.
  • Responsible to manage data coming from different sources

Confidential

Environment: HBASE, Sqoop, MapReduce, HDFS, UDFs,Java, JSP, Oracle 11g and PL/SQL

Team Leader

Responsibilities:

  • Involved in scoping the application requirement and functional specifications
  • Designed the Database for this application
  • Involved in developing the complex queries from Oracle
  • Optimized system performance by writing UDF’s and stored procedures
  • Involved in production support
  • Used SVN as the version Control system
  • Developed multi data source based transaction process logic
  • Assisted new developers on development and architecture issues
  • Involved in conduction code review and design review for junior developers.
  • Installed and configured Hadoop on a cluster
  • Experienced in defining job flows using Oozie
  • Experienced in managing and reviewing the Hadoop log files
  • Load and transform large sets of structured (using Sqoop), semi structured (using M/R) and unstructured (using M/R) data.
  • Responsible to manage data coming from different sources and application.
  • Loading the data into MSTR (Micro Strategy) cube.
  • Involved in Unit level testing
  • Prepared design documents and functional documents.
  • Connected local file system to HDFS using WinScp.
  • Based on the requirements, addition of extra nodes to the cluster to make it scalable.
  • Check the memory usage and disk usage to find the malfunctioning nodes.
  • Submit a detailed report about the daily activities on a weekly basis

Confidential, New Jersey

Environment: Oracle 11g, Java, JSP, PL/SQL & Teradata

Team Leader

Responsibilities:

  • DB design, code review, performance tuning, implemented critical procedures and EOD’s.
  • Migrating required data from existing DB to new Data Center.
  • New affiliate registration (VRF-Vendor Registration Form) and submitting the same to DGE.
  • Implementing the new Invoicing, new Taxation and new Cashier process for Affiliates
  • Working as Scrum Master for this project.

Confidential

Environment: HIVE, Sqoop, MapReduce, UDFs, Java, JSP, Oracle 11g and PL/SQL

Team Leader

Responsibilities:

  • Involved in scoping the application requirement and functional specifications
  • Designed the Database and the business layer for this application
  • Involved in developing the complex queries from Oracle, Hive and Sqoop
  • Optimized system performance by writing Hive UDF’s and stored procedures
  • Involved in production support
  • Used SVN as the version Control system
  • Assisted new developers on development and architecture issues
  • Involved in conduction code review and design review for junior developers.
  • Installed and configured Hadoop on a cluster
  • Experienced in defining job flows using Oozie
  • Experienced in managing and reviewing the Hadoop log files
  • Load and transform large sets of structured (using Sqoop), semi structured (using M/R) and unstructured (using M/R) data.
  • Responsible to manage data coming from different sources and application.
  • Installed and configured Hive and also written Hive UDFs.
  • Involved in creating Hive tables, loading with data and writing hive queries which will run internally in map reduce way.
  • Exposing the Hive tables to the external world to for loading and KPI reports purpose
  • Loading the data into MSTR (Micro Strategy) cube.
  • Involved in Unit level testing
  • Prepared design documents and functional documents.
  • Connected local file system to HDFS using WinScp.
  • Based on the requirements, addition of extra nodes to the cluster to make it scalable.
  • Check the memory usage and disk usage to find the malfunctioning nodes.
  • Submit a detailed report about the daily activities on a weekly basis

Confidential

Environment: Oracle 11g, Java, JSP, PL/SQL, PL/SQL Web Cartridge, Teradata, Informatica, PSP, BO

Team Leader

Responsibilities:

  • DB Design, Code review, Tuning, Report development and loading & processing the data into various DB’s.
  • Migrating all existing deals and other key information as per grouped brands
  • Integrating with other supporting systems DB’s like Cashier, Gaming and Data warehouse…
  • Re-branding changes should not affect retrospectively for all analysis and KPI reports from DWH
  • Commission calculating as per new business rules
  • Changing the ETL (Informatica) mappings accordingly and loading the data in DWH (Teradata)

Confidential

Environment: Oracle 11g, PL/SQL, Teradata, Informatica, MSTR (Micro Strategy)

Team Leader

Responsibilities:

  • DB Design, Code review, Tuning, Report development and loading & processing the data into various DB’s.
  • Summarizing all the data from various entity at player, game and country level
  • Calculating the commission at player level for affiliates
  • Moving data from OLTP (ORACLE) to OLAP (Teradata) using ETL (Informatica)and processing.
  • Moving the data from Teradata to MSTR(Micro Strategy) CUBE
  • Providing CUBE reports to International Accounting team and Affiliates Marketing team
  • Loading the data on daily / weekly / monthly basis

Confidential

Environment: Oracle 10g, Teradata, PL/SQL, PL/SQL Web Cartridge, Informatica, PSP, BO

Team Leader

Responsibilities:

  • Migration of Confidential .com affiliates and player system into party gaming affiliates and player system
  • Filling the functional gaps between Confidential and PARTY
  • De-commissioning the applications, without changing much in the Database structures.
  • New player Qualification system as per new business criteria
  • Loading the required data into Teradata using Informatica
  • Make available all the data to Business layer for KPI reporting and all

We'd love your feedback!