We provide IT Staff Augmentation Services!

Hadoop Developer/lead Resume

0/5 (Submit Your Rating)

Bentonville, AR

SUMMARY

  • 7+ years of experience in designing and developing Business solutions.
  • Skilled in Planning, Designing, developing and deploying Data warehouses.
  • Strong experience in designing and developing Business solutions in Data Warehouse/Decision Support Systems using Pivotal version Hadoop, Hive, PIG, Sqoop, Big SQL, TDCH and Spark.
  • Experience in integration of various data sources with Multiple Relational Databases like Oracle, Teradata, Green plum, SAP HANA and Worked on integrating data from flat files like fixed width and delimited.
  • Experience in Designing and deploying in Big Data Hadoop ecosystem in Cloudera.
  • Expert knowledge on HIVE batch processing techniques and HIVE data units including partitioning and bucketing.
  • Experience in designing framework using ETL tool Syncsort.
  • Exposure to Apache HIVE file formats like ORC, AVRO, JSON and Sequence Files
  • Worked with sources and Confidential such as XML, COBOL file, MQ sources/Targets, SAP, Flat files (Delimited and fixed width), Oracle, Sql Server, and Teradata etc.
  • Used HCatalog over HIVE with Pig Latin to process input/output of batch data.
  • Strong understanding of Mapper, Combiner and Reducer processes in Hadoop YARN cluster.
  • Data sources worked on include Flat Files, Excel files and Databases such as Green plum, Teradata and SAP HANA.
  • Programmed in Python for over four years and familiar with functional programming concepts.
  • Improved performance by using Explain Plan, Creating appropriate indexes, queries optimization using incremental load /daily refresh and using TEZ, ORC (Optimized row columnar), Partition tables and Parallel execution in Hive.
  • Experience in Healthcare, Banking domains and Retail Industry.
  • Experienced in interacting with business users, business analysts.
  • IT leads and developers in analyzing business requirements and translating requirements into functional and technical design specifications.
  • Excellent communication, presentation and interpersonal skills and ability to prioritize and coordinate work across different geographic locations.
  • Experience working with agile methodology.

TECHNICAL SKILLS

Operating Systems: UNIX, Z/OS, OMVS, Windows

Scripting Languages: Unix

Databases: Sql Server, MySQL, Big SQL, Teradata, Greenplum, SAP HANA, DB2

Languages: SQL, PL/SQL, Python, COBOL, JCL

Technologies: Sqoop 1.4.5, Oozie 4.1.0.3.0.0.0 - 249 , Hadoop 2.6.0.3.0.0.0-249 , Hive 0.14.0.3.0.0.0-249 , Pig 0.14.0.3.0.0.0-249 , Spark 1.3.1

Database Tools: SQL Developer, SAP HANA Studio, SyncSort, Teradata SQL assitant, Pgadmin Tool.

PROFESSIONAL EXPERIENCE

Confidential, Bentonville, AR

Hadoop Developer/Lead

RESPONSIBILIES:

  • Worked with Application Team to ingest Supply chain Data object.
  • Experience in designing framework/Data Modelling using Syncsort.
  • Collaborated with Business Analysts to ascertain the issues with the existing data warehouse. Modified mappings to conform to business rules.
  • Experience in Cloudera Hadoop for executing Business requirements using HIVE, PIG, SQOOP and Spark.
  • Develop Aggregation logic using Hive and Spark.
  • Developed shell scripts to perform the testing.
  • Developed Load script to Green plum using Python, PostgreSQL.
  • Responsible for creating the parameter files in order to connect to the right environment and DB.
  • Worked with DBA on SQL scripts to automate the process of populating the various columns in the tables.
  • Provided production support.
  • Conceived unit, system, integration, functional, and performance test plans.

Environment: Hadoop 2.6.0.3.0.0.0-249 , HDFS, Hive 0.14.0.3.0.0.0-249 , 249, Pig 0.14.0.3.0.0.0-249 , Spark 1.3.1, Oozie 4.1.0.3.0.0.0-249 , Sqoop 1.4.5, Unix Shell scripting, CA7 Scheduler, Big SQL, Teradata, Green plum and SAP HANA and SDLC.

Confidential, Bentonville, AR

Hadoop Developer

RESPONSIBILIES:

  • Worked in Customer Knowledge Platform (CKP) to provide data about the customer.
  • Developed aggregations and metrics on customer transactions and interactions for enabling GCIA.
  • Converted historical data into the new streamlined standards. Created new data objects for different business needs.
  • Collaborated with Business Analysts to ascertain the issues with the existing data warehouse. Modified mappings to conform to business rules.
  • Developed Python application and shell scripts to perform the testing.
  • Worked with DBA on SQL scripts to automate the process of populating the various columns in the tables.
  • Provided production support.
  • Conceived unit, system, integration, functional, and performance test plans.

Environment: Hadoop 2.6.0.3.0.0.0-249 , HDFS, Hive 0.14.0.3.0.0.0-249 , 249, Pig 0.14.0.3.0.0.0-249 , Spark 1.3.1, Oozie 4.1.0.3.0.0.0-249 , Sqoop 1.4.5, Unix Shell scripting, CA7 Scheduler, Teradata, Green plum and SAP HANA.

Confidential, Minneapolis, MN

Hadoop Developer

RESPONSIBILIES:

  • Documented user requirements, translated requirements into system solutions and developed implementation plan and schedule.
  • Monitoring and supporting applications stability using COBOL, DB2, JCL, and Control M,Hadoop, Hive.
  • Managed writing test cases and test scenarios from requirement for newly added features and executing test scripts.
  • Developed Load script to capture sales information in Hadoop using Hive.
  • Identifying and performing the break fixes in the production applications.
  • Develop quality code adhering to Python coding Standards and best practices.
  • Taking care of Quality Assurance activities for all deliverables.
  • Organized data in the report Inserting Filters, Sorting, Ranking and highlighting data for Vendor returns and unsalable application.
  • Provided production support.
  • Conceived unit, system, integration, functional, and performance test plans.

Environment: Unix, IBM zOS/390, COBOL, Easytrieve, Xpeditor, Endeavor, DB2, Hadoop 2.3.0, HDFS/Hive.

Confidential, Minneapolis, MN

Hadoop Developer

RESPONSIBILIES:

  • Documented user requirements, translated requirements into system solutions and developed implementation plan and schedule.
  • Monitoring and supporting applications stability using COBOL, DB2, JCL, and Control M, Hadoop, Hive, Sqoop.
  • Implementeddesign patternsin Python for the application.
  • Developed Load script to replicate the logic of the process running in Teradata to Hadoop.
  • Identifying and performing the break fixes in the production applications.
  • Taking care of Quality Assurance activities for all deliverables.
  • Provided production support.
  • Conceived unit, system, integration, functional, and performance test plans.

Environment: Unix, IBM zOS/390, COBOL, Easytrieve, Xpeditor, Endeavor, DB2, Hadoop 2.3.0, HDFS/Hive.

Confidential, New Jersey

Developer

RESPONSIBILIES:

  • Interacted with End Users for gathering requirements.
  • Assessed the information and business needs of the users and mapped Business requirements to Technical terms.
  • Developed Technical Design Documents for MSSB Conversion.
  • Developed new programs to execute the business requirement.
  • Performed Unit Testing, Volume Testing and Regression testing and also prepared Unit Test Plan (UTP), Unit Test Results (UTR), System Test plan (STP) and System Test Results (STR) as part of testing activities.
  • Created Validation reports for Business users to validate the account conversion.

Environment: IBM ZOS/390, COBOL, Changeman, Xpeditor, FILEAID and DB2

We'd love your feedback!