We provide IT Staff Augmentation Services!

Hadoop Developer/technical Lead/manager Resume

3.00/5 (Submit Your Rating)

SUMMARY

  • Seasoned professional with 13+ Years of experience, 2+ years of experience in Apache Hadoop, Big Data analytics, in application development, support and team management.
  • Vast experience in Agile Methodologies and implementing this in Retail and Automotive applications.
  • Strong experience in working on diverse technical platforms and Proficient understanding of distributed computing principles.
  • Experience in Hadoop development, implementation in Cloudera distribution.
  • Exclusive experience in Hadoop Ecosystem and its components like HDFS, Apache Pig, Hive, Sqoop, Flume, Kafka, Spark.
  • Having experience with processing real time streamed data using Spark streaming, Spark SQL.
  • Experienced in transporting, and processing real time event streaming using Kafka and Spark Streaming.
  • Comprehensive experience in coding for the following Hadoop Eco System - Spark/Python, Spark/RDD, Spark SQL.
  • Strong experience in pre-processing and Data querying tools, such as Pig, Hive.
  • Knowledge of NoSQL databases -MongoDB.
  • Strong experience with integration of data from multiple data sources
  • Proficient in relational databases and SQL Database design, creation, manipulation and performance tuning.
  • Extensive experience in data modelling technique and Hands on experience with Oracle databases
  • Hands on experience with Unix Shell Scripting.
  • Highly experienced in creating Lower/Higher Level Design, Requirement Specification, and Impact Analysis documents.
  • Strong understanding of Cobol DB2 Stored Procedures and working knowledge of SQL Server.
  • Good working knowledge of Relational Databases like IBM DB2, SQL server. I can easily work on any RDBMS.
  • Proficient in in implementing and managing applications using TSO/ISPF, COBOL, DB2, JCL, Procs, SQL, Rexx, EasyTrieve, DYL280.
  • Extensive experience in various types of testing - Unit Testing, Regression Testing, Integration and System Testing and implementation of Shift Left Approach.
  • Excellent in team management and working with onsite-offsite work model.
  • Strong experience working with issue tracking and change management tools like IBM Maximo, Service Desk Manager.
  • Quick in adapting new software’s and promised to deliver quality work to business.
  • Proficient in customer management and skilled in working with large teams under challenging and dynamic environment.
  • Pursue excellent Oral and written communication skills.

TECHNICAL SKILLS

Ingestion Tools: Flume, Sqoop, Kafka

Databases: Mongo DB

Data Transformation and Framework: Pig, Hive, Python, PySpark, Spark/RDD, SparkSQL

Shell Scripting: Mainframe Platform

Languages: Cobol, JCL, SQL, Rexx, CICS, IMS DC, EasyTrieve, DYL280

Databases: IBM DB2, IMS DB

Utilities: TSO/ISPF, File Aid, File Manager, VSAM, SPUFI, QMF, Xpeditor, SYNCSORT, JOBTRAC, Platinum, SAR and CA-7

Windows Batch Scripting: SQL Server

Tools: JIRA, IBM Manage Now, IBM Maximo, CA Service Desk Manager

PROFESSIONAL EXPERIENCE

Confidential

Hadoop Developer/Technical Lead/Manager

Responsibilities:

  • Created Hive tables, and loading and analyzing data using hive queries.
  • Worked on debugging, performance tuning of Hive & Pig Jobs.
  • Writing complex Python code to process data in Hadoop cluster.
  • Ingesting real time data from Source Kafka into HDFS using Spark Streaming.
  • Designed and implemented pig UDFs for evaluation, filtering, loading and storing of data.
  • Worked on Performance Tuning of Hadoop jobs by applying techniques such as MapSide Joins, Partitioning, Bucketing and using different file formats such as SequenceFile, RCFile, ORC File, Parquet File, Avro file.
  • Proactively monitored systems and services, architecture design and implementation of Hadoop deployment, configuration management, backup, and disaster recovery systems and procedures.
  • Used Kafka to collect, aggregate, and store the web log data from different sources like web servers, mobile and network devices and pushed to HDFS.
  • Convert Spark RDD to Data Frames.
  • Load and transform large sets of structured, semi structured and unstructured data
  • Ingesting through Kafka and processing through Spark Streaming.
  • Importing and exporting data into HDFS and Hive using Sqoop, Spark Core and Spark SQL, Python.
  • Involved in loading data from UNIX file system to HDFS, configuring Hive and writing Hive UDFs.
  • Involved in writing application code, testing and Deployment of code.
  • Worked in converting Hive/SQL queries into Spark transformations using Spark RDDs and Python.
  • Developed Python and SQL code to extract data from various databases.
  • Developed Spark code using Python and Spark-SQL for faster testing and data processing.
  • Managed and reviewed log files
  • Development and maintenance of new components using Cobol, DB2, SQL and enhancements and testing of the existing ones using JCL, Xpeditor, SPUFI.
  • Optimizing and tuning SQLs based on Explain Plan, to improvise the system performance.
  • Meeting with Business for encapsulating requirements for developments and enhancements to deliver scalable and extensible software solutions.
  • Preparation of Lower Level Design Document, Approach Documents, Scope Document and Pseudo code Document
  • Working with business to closely work on Impact Analysis and feasibility study for any new requirement.
  • Monitoring and scheduling batch jobs using Jobtrac.
  • Providing Level 1 production support and work closely with Operation Analyst to resolve abends within time to provide smooth run of batch cycles.
  • Performing extensive Unit testing, System testing, Regression testing and User acceptance testing to maintain the consistent and high quality of deliverables.

Environment: Hive, Flume, Sqoop, Pig, Kafka, MangoDB, Oozie, SQL Server, Cobol DB2, SQL, QMF, SPUFI, Endeavor, File-Manager, DB2 Platinum, SQL server.

Confidential

Analyst Programmer 2

Responsibilities:

  • Requirement gathering from clients and passing the abstract to offshore team members in form of Application package (AP) and Development package (DP).
  • Involved in Physical data Modeling and creating lower level design documents based on the business requirement.
  • Extensively worked on Cobol DB2 Stored Procedures at the back end and Java Interface at the back end to implement Dealer Maintenance Screen.
  • Worked on creating online screens using IMS DC for front end and IMS-DB as the database on the back end.
  • Coding of new COBOL DB2, SQL, IMS DB- IMS-DC programs and maintaining the different versions using Panvalet as the Version control tool.
  • Was a member of Performance Tuning Team and involved in analysis of queries using Plan Analyzer output such as cardinality, filter factor, Index able Columns and CPU time.
  • Involved in change of existing JCL, Procs, Copybooks, Includes and their testing using debugger as Xpeditor afterwards.
  • Involved in pear to peer review of components delivered by offshore team members to ensure the consistent good quality of deliverables.
  • Involved in preparation of Unit test plan and unit test results in various phases of unit testing, Integration testing and System testing.
  • Estimation of tasks and subtasks based on resources and priority of work.
  • Allocation of tasks to the team members
  • Development and testing of new programs
  • Modification and Enhancement of Programs
  • Change of the existing JCL/procedures/includes/copybooks
  • Unit testing and preparation of test plan, test scripts, test cases.
  • System testing by embedding new/modified code to the system.
  • Involved in Delivery Co-Ordination between Onsite & Offshore Teams
  • Following the Project Development Life Cycle for each task

Environment: TSO/ISPF, COBOL, DB2, SQL, JCL, Stored procedures, Procs, CA-7, Panvalet, File-Aid, DB2 Platinum, SPUFI, QMF, Xpeditor.

We'd love your feedback!