We provide IT Staff Augmentation Services!

Hadoop/spark Developer Resume

3.00/5 (Submit Your Rating)

Richmond, VA

SUMMARY:

  • Over 11 years of experience, including 3 years on Big Data Hadoop Eco Systems with hands - on project experience in health care domain.
  • Cloudera certified Spark and Hadoop developer
  • Extensive experience on SQL&PL/SQL, Hadoop and Linux Shell Scripting
  • Expertise in developing Spark applications using Scala
  • Exposure to data quality analysis, data migration and data validation.
  • Migrating the coding from Hive to Apache Spark and Scala using SparkSQL, RDD
  • Working experience on Spark, Spark SQL, RDD, Data Frames and Datasets
  • Hands on experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems and vice-versa.
  • Good Experience in file handling operations on Unix.
  • Professionally adaptable to 'get on board' quickly with new technologies and new assignments, able to work independently or in a group/team setting to meet deadlines while handling multiple assignments.
  • Extensively involved in Requirement analysis, Code development and Bug fixing.
  • Expertise in Hive Query Language and debugging Hive issues.
  • Expertise in Sqoop and Flume
  • Extremely motivated with good inter-personal Skills have ability to work in strict deadlines
  • Accomplished facilitator in understanding
  • Proficiency in SDLC Life Cycle, Agile and Waterfall models
  • Coordinating with offshore team on technical requirements, design, test plan, test results, pre-implementation activities and deployment

TECHNICAL SKILLS:

Big data ecosystem: Hadoop, MapReduce, HDFS, Hive, Flume, Kafka, Sqoop, Hbase, Zookeeper, Oozie, Apache Spark, SparkSQL, Spark Streaming, RDD, Data Frames, JSON, AVRO

Programming/scripting languages: Scala, Linux shell scripts, Cobol, JCL, XML

Databases: IBM-DB2, MySQL, IDMS, IMS-DB

Operating systems: Unix, centos 6.7, Z/OS, MVS, Windows …

Development tools: SUBLIME 2, Eclipse, AVRO-TOOLS

Collaboration Tools: JIRA, Rational Team Concert, Rational ClearQuest, SharePoint

PROFESSIONAL EXPERIENCE:

Hadoop/Spark Developer

Confidential, Richmond, VA

Responsibilities:

  • Involved in developing roadmap for migration from legacy system to Hadoop cluster.
  • Create, validate and maintain scripts to load data using Sqoop manually.
  • Load and transform large sets of structured, semi structured and unstructured data coming from various downstream systems.
  • Migrated data between RDBMS and HDFS/Hive with Sqoop.
  • Create, validate and maintain scripts to extract and transform data from MySQL to flat files and JSON format.
  • Create Oozie workflows and coordinators to automate Sqoop jobs weekly and monthly.
  • Worked on reading multiple data formats on HDFS using Apache Spark.
  • Wrote Scala scripts for spark to perform operations like data inspection, cleaning, load and transforms the large sets of structured and semi-structured imported data
  • Involving in Migrating the coding from Hive to Apache Spark and Scala using SparkSQL, RDD
  • Developed Spark with Scala and Spark sql for testing and processing of data
  • Developed, validate and maintain HiveQL queries.
  • Designed Hive tables to load data to and from external HDFS datasets
  • Hands on using Partitions, Bucketing concepts in Hive and designed both Managed and External tables in Hive for optimized performance.
  • Managing and scheduling jobs on a Hadoop cluster.

Environment: Hadoop, HDFS, Apache Spark 1.6,SparkSQL, Unix, Hive, Sqoop, Flume, Scala, Oozie, DB2 SQL, kafka

Senior System Analyst

Confidential, Albany, NY

Responsibilities:

  • As Senior System Analyst, managed and coordinated multiple systems/teams involved to ensure successful delivery of the efforts as per the Business Requirements
  • Ensure on-time delivery for the SDLC phases for projects and SSCRs (Small System Change Requests)
  • Perform weekly work forecast, LOE reviews and related negotiations
  • Represent the CHIPS VA provider and CS90 NY provider teams in the weekly Client Portfolio and Stake Management calls
  • Identify and raise Risks, Issues and dependencies to the stakeholders and come up with mitigations plans in a timely manner
  • Review the deployment plan and document implementation strategies with the Confidential management and SME's. Discuss risks, issues and dependencies in the implementation procedures in a deployment across applications
  • Prepare and review Level of Estimate in hours for each of the efforts executed at Confidential . Based on industry standards the LOE is prepared with clear breakup for each phase of the project and sometimes at activity level
  • Arrange internal JAD sessions and come up with the best possible design approach to the many projects and SSCRs
  • Interacting with SME's, Business, Application owners/end users to determine application requirements, user problems, participating in architecture and design activities
  • SPOC for the CHIPS VA provider and CS90 NY provider team for the release management activities for the Confidential CHIPS Virginia System and CS90 Newyork systems.
  • Owner of the Procedures and Process documentation for the Confidential account and its maintenance

Environment: Z/OS, MVS, COBOL, JCL/PROC, VSAM, Cobol-XML, DB2 SQL, SPUFI, IMS, REXX, CLIST, VB scripting, ClearQuest, Share-point

Batch Process Lead & Application developer

Confidential

Responsibilities:

  • Studying the different systems that exist in Health-care insurance and their functional impact.
  • Understanding the client requirements and preparation of technical documents.
  • Working and implementing the code in the mainframe environment.
  • Understanding the interfaces for claims software and PLASM language.
  • Maintenance, enhancement and developments of programs coded in COBOL as per the requirements.
  • Working on the Incidents primarily concerning claims (maintenance support system).
  • Supporting the high-priority live batch flow system.

Environment: Z/OS, MVS, COBOL, IMS-DB/DC, IDMS, JCL/PROC, REXX, CLIST.

We'd love your feedback!