We provide IT Staff Augmentation Services!

Senior Hadoop Developer Resume

4.00/5 (Submit Your Rating)

New, YorK

PROFILE:

  • 9 Years of IT industry experience encompassing a wide range of skill sets. Roles and industry verticals.
  • Certified Big data (HADOOP) developer, Certified in IBM DB2 730/733 Developer and lean concepts
  • Hands on experience in performing development and data analytics using HADOOP (BIG - DATA) tools and technologies which included HDFS, MAP REDUCE, HIVE, PIG, HBASE, SPARK, FLUME, SQOOP and OOZIE.
  • Strong database skills in DB2, HIVE, Oracle, PL/SQL, MySQL, BigSQL and No-SQL databases like HBASE, familiarity with CASSANDRA.
  • Experienced in installing, configuring, managing, and testing Big-data “HADOOP” ecosystem components.
  • Experienced in developing map reduce program using java.
  • Used Apache Spark for large-scale data processing, handling real-time analytics and real streaming of data
  • Experienced in Data warehouse concepts and ETL tools (Teradata).
  • Experienced using Teradata SQL Assistant, data import/export, data loading with utilities like BTEQ, Multi Load, Fast Load, and Fast Export on UNIX/Mainframes environments.
  • Experienced in Stored Procedure, Trigger and macros, SQL Loader.
  • Experienced in UNIX Shell Scripting
  • Experience in the area of legacy systems based on Mainframe platform
  • Good knowledge in Maestro, StartTeam, Buildforge.
  • Experienced with workflow schedulers, data architecture including data ingestion pipeline design and data modeling.
  • Possess functional knowledge in the areas of Insurance systems, Financial Systems, Banking System and Healthcare System.
  • Good experience in all phases of systems life cycle Development, Testing (Unit test, System test, Integration Testing and Regression Testing) and Pre-Production support.
  • Proficient in analyzing and translating business requirements to technical requirements and architecture.
  • Performed Knowledge management in the form of AIDs and Project knowledge and change documents.
  • Experienced in handling internal and external functional, process and data audits.

TECHNICAL SKILLS:

Big Data Ecosystems: HDFS, Hive, Pig, Map Reduce, Spark Sqoop, HBase, Cassandra, Zookeeper, Flume, Oozie, Avaro and Hue

Languages: Java, PL/SQL, Python, Scala Unix shell scripting, Hiveql, Pig scripts, and Cobol

Data Base: MY SQL, BIGSQL, NOSQL, Oracle, DMS1100 and DB2

Operating System: Unix, Windows, MVS/ESA, ZOS

ETL/Reporting: Teradata

Methodologies: Waterfall, Scrum, and Agile

Tools: RPM, MPP, Test Direct, Clarity, Quality Center, Service Center, SFTP, Teradata Sql assistant, Toda, SSH, HUE, Eclipse, Maven, Putty, BigInsight, Cloudera

PROFESSIONAL EXPERIENCE:

Confidential, New York

Senior Hadoop Developer

Responsibilities:

  • Create the project using HIVE, IMPALA, PIG and SPARK
  • Implemented Partitioning, Bucketing in HIVE.
  • Involved in data modeling in Hadoop.
  • Creating Hive tables and working on them using Hiveql.
  • Written Apache PIG scripts to process the HDFS data.
  • Created Java UDFs in PIG and HIVE.
  • Experience in using Sequence files, AVRO, SAN and ORC file formats.
  • Involved in data modeling in Hadoop.
  • Good working knowledge of Amazon Web Service components like EC2, EMR, S3, EBS, ELB
  • Come up with estimations and Technical Design Specifications for projects.
  • Requirement Analysis & Prepares solutions for each requirement
  • Gathered the business requirements from the Business Partners and Subject Matter Experts.

Environment: HADOOP, HDFS, MAPREDUCE, HIVE, PIG, Scala, HBASE, OOZIE, yarn, Spark, Core Java, Teradata, SQL, UBUNTU/UNIX, eclipse, Maven, JDBC drivers, Mainframe, MySQL, Linux, AWS, XML, CRM, SVN, HUE, Putty, Cloudera

Confidential, CA

Senior Hadoop Developer

Responsibilities:

  • Create the project using HIVE, BIGSQL, PIG
  • Implemented Partitioning, Bucketing in HIVE.
  • Involved in data modeling in Hadoop.
  • Creating Hive tables and working on them using Hiveql.
  • Written Apache PIG scripts to process the HDFS data.
  • Created Java UDFs in PIG and HIVE.
  • Experience in using Sequence files, RCFile, AVRO and ORC file formats.
  • Involved in data modeling in Hadoop.
  • Participated in backup and recovery of Hadoop file system.
  • Automated tasks using UNIX shell scripts.
  • Requirement Analysis & Prepares solutions for each requirement
  • Gathered the business requirements from the Business Partners and Subject Matter Experts.

Environment: HADOOP, HDFS, MAPREDUCE, HIVE, PIG, Scala, Python, HBASE, OOZIE, yarn, Spark, Core Java, Oracle, SQL, UBUNTU/UNIX, eclipse, Maven, JDBC drivers, Mainframe, MySQL, Linux, AWS, XML, CRM, SVN, PDSH, Putty, BigInsights

Confidential, NJ

Senior Developer

Responsibilities:

  • Understand the requirement and build the HBASE data model
  • Loaded history Data as well as incremental customer and other data to Hadoop through Hive.
  • Applied the required Business logic to the data in hive and generated the required output in the form of Flat file.
  • Experienced in writing complex Pig jobs.
  • Importing and exporting large data sets from various data sources into HDFS using Sqoop.
  • Implemented Partitioning, Bucketing in HIVE.
  • Load balancing of data across the cluster and performance tuning of various jobs running on the cluster.
  • Involved in analyzing and debugging errors occurring during jobs execution in Big Data cluster environment.
  • Developed Oozie workflow for scheduling and orchestrating the ETL process.
  • Provide solutions to Walkups, operational, incident tickets
  • Provide data fixes and code fixes related to defects
  • Developed Queries for reporting
  • Developed applications using Eclipse
  • Performed process enhancement by SQL Tuning.
  • Provide low level and high level solution design document.
  • Responsible for disaster recovery of systems
  • Participate and Perform software upgrades and conversions.
  • Translate customer requirements into formal requirements and design documents, establish specific solutions, and leading the efforts including programming and testing that culminate in client acceptance of the results.

Environment: HADOOP, HDFS, MAPREDUCE, java, HIVE, Hue, PIG, Flume, SQOOP, HBASE, OOZIE, Yarn, Zookeeper eclipse, Maven, BigInsight

Confidential, NJ

Senior Developer

Responsibilities:

  • Requirement Analysis & Prepares solutions for each requirement
  • Designed TDD (low level) from SRS (High level)
  • Assigning of task, daily updates, weekly status update to client.
  • Responsible for Design, Data Mapping Analysis and Mapping rules
  • Used Python script to transform the data.
  • Fixed issues with the existing Fast Load/ Multi Load Scripts in for smooth loading of data in the warehouse more effectively.
  • Worked on loading of data from several flat files sources to Staging using MLOAD, FLOAD.
  • Created Bteq scripts with data transformations for loading the base tables.
  • Generated reports using Teradata BTEQ.
  • Worked on optimizing and tuning the Teradata SQLs to improve the performance of batch and response time of data for users.
  • Fast Export utility to extract large volume of data and send files to downstream applications
  • Created stored procedure as per business requirement and involved in performance tuning

Environment: Teradata V2R12, Teradata SQL Assistant, MLOAD, FASTLOAD, BTEQ, Erwin, Unix Shell Scripting, Macros, Stored procedure, Db2, Cobol, Python, SAS, PL/SQL, FileZilla

Confidential

Developer

Responsibilities:

  • Created and reorganized all types of database objects including tables, views, indexes, sequences, synonyms and setting proper parameters and values for all the objects.
  • Wrote database triggers, stored procedures, stored functions, and stored packages to perform various automated tasks for better performance.
  • Created indexes on the tables for faster retrieval of the data to enhance database performance.
  • Created Shell Scripts for invoking SQL scripts.
  • Created and modified several UNIX shell scripts according to the changing needs of the project.
  • Used different joins, sub-queries and nested queries in SQL query
  • Effectively made use of Table Functions, Indexes, Table Partitioning, Analytical functions, and Materialized Views
  • Used the Cursor for Loop to fetch unconditional number of rows
  • Imported/Exported data from/to different databases using utilities like SQL*Loader.
  • Experience with Performance Tuning for Oracle RDBMS using Explain Plan and HINTS.
  • Involved in the continuous enhancements and fixing of production problems.
  • Verified and validated data using SQL queries.
  • Analyzed and prepared High and low level designs.
  • Made clear, maintainable, efficient and reusable codes
  • Provide post-production support to Developed modules during QA&UAT phase

Environment: Oracle 10g, java, SQL, PL/SQL, UNIX, SQL*Loader, SQL Navigator, TOAD, SQL DEVELOPER.

We'd love your feedback!