Hadoop/spark Developer Resume
Richmond, VA
SUMMARY:
- Over 11 years of experience, including 3 years on Big Data Hadoop Eco Systems with hands - on project experience in health care domain
- Cloudera certified Spark and Hadoop developer
- Extensive experience on SQL&PL/SQL, Hadoop and Linux Shell Scripting
- Expertise in developing Spark applications using Scala
- Exposure to data quality analysis, data migration and data validation.
- Migrating the coding from Hive to Apache Spark and Scala using SparkSQL, RDD
- Working experience on Spark, Spark SQL, RDD, Data Frames and Datasets
- Hands on experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems and vice-versa.
- Good Experience in file handling operations on Unix.
- Professionally adaptable to 'get on board' quickly with new technologies and new assignments, able to work independently or in a group/team setting to meet deadlines while handling multiple assignments.
- Extensively involved in Requirement analysis, Code development and Bug fixing.
- Expertise in Hive Query Language and debugging Hive issues.
- Expertise in Sqoop and Flume
- Extremely motivated with good inter-personal Skills have ability to work in strict deadlines
- Accomplished facilitator in understanding
- Proficiency in SDLC Life Cycle, Agile and Waterfall models
- Healthcare domain experience of close to 10 years with detailed knowledge on Membership Enrollment, Claims Adjudication process and provider systems
- Coordinating with offshore team on technical requirements, design, test plan, test results, pre-implementation activities and deployment
TECHNICAL SKILLS:
Big data ecosystem: Hadoop, MapReduce, HDFS, Hive, Flume, Kafka, Sqoop, Hbase, Zookeeper, Oozie, Apache Spark, SparkSQL, Spark Streaming, RDD, Data Frames, JSON, AVRO
Programming/scripting languages: Scala, Linux shell scripts, Cobol, JCL, XML
Databases: IBM-DB2, MySQL, IDMS, IMS-DB
Operating systems: Unix, centos 6.7, Z/OS, MVS, Windows
Development tools: SUBLIME 2, Eclipse, AVRO-TOOLS
Collaboration Tools: JIRA, Rational Team Concert, Rational ClearQuest, SharePoint
PROFESSIONAL EXPERIENCE:
Confidential - Richmond, VA
Hadoop/Spark Developer
Responsibilities:
- Involved in developing roadmap for migration from legacy system to Hadoop cluster.
- Create, validate and maintain scripts to load data using Sqoop manually.
- Load and transform large sets of structured, semi structured and unstructured data coming from various downstream systems.
- Migrated data between RDBMS and HDFS/Hive with Sqoop.
- Create, validate and maintain scripts to extract and transform data from MySQL to flat files and JSON format.
- Create Oozie workflows and coordinators to automate Sqoop jobs weekly and monthly. Worked on reading multiple data formats on HDFS using Apache Spark.
- Wrote Scala scripts for spark to perform operations like data inspection, cleaning, load and transformsthe large sets of structured and semi-structured imported data
- Involving in Migrating the coding from Hive to Apache Spark and Scala using SparkSQL, RDD
- Developed Spark with Scala and Spark sql for testing and processing of data
- Developed, validate and maintain HiveQL queries.
- Designed Hive tables to load data to and from external HDFS datasets
- Hands on using Partitions, Bucketing concepts in Hive and designed both Managed and External tables in Hive for optimized performance.
- Managing and scheduling jobs on a Hadoop cluster.
Environment: Hadoop, HDFS, Apache Spark 1.6,SparkSQL, Unix, Hive, Sqoop, Flume, Scala, Oozie, DB2 SQL, kafka
Confidential
Technical Team lead and System Analyst
Responsibilities:
- As Delivery Lead, managed and coordinated multiple systems/teams involved to ensure successful delivery of the efforts as per the Business Requirements
- Ensure on-time delivery for the SDLC phases for projects and SSCRs (Small System Change Requests)
- Perform weekly work forecast, LOE reviews and related negotiations
- Represent the CHIPS VA provider and CS90 NY provider teams in the weekly Client Portfolio and Stake Management calls
- Identify and raise Risks, Issues and dependencies to the stakeholders and come up with mitigations plans in a timely manner
- Review the deployment plan and document implementation strategies with the Anthem management and SME's. Discuss risks, issues and dependencies in the implementation procedures in a deployment across applications
- Prepare and review Level of Estimate in hours for each of the efforts executed at Anthem. Based on industry standards the LOE is prepared with clear breakup for each phase of the project and sometimes at activity level
- Arrange internal JAD sessions and come up with the best possible design approach to the many projects and SSCRs
- Interacting with SME's, Business, Application owners/end users to determine application requirements, user problems, participating in architecture and design activities
- SPOC for the CHIPS VA provider and CS90 NY provider team for the release management activities for the Anthem CHIPS Virginia System and CS90 Newyork systems.
- Owner of the Procedures and Process documentation for the Anthem account and its maintenance
Environment: Z/OS, MVS, COBOL, JCL/PROC, VSAM, Cobol-XML, DB2 SQL, SPUFI, IMS, REXX, CLIST, VB scripting, ClearQuest, Share-point
Confidential
Batch Process Lead & Application developer
Responsibilities:
- Studying the different systems that exist in Health-care insurance and their functional impact.
- Understanding the client requirements and preparation of technical documents.
- Working and implementing the code in the mainframe environment.
- Understanding the interfaces for claims software and PLASM language.
- Maintenance, enhancement and developments of programs coded in COBOL as per the requirements.
- Working on the Incidents primarily concerning claims (maintenance support system).
- Supporting the high-priority live batch flow system.
Environment: Z/OS, MVS, COBOL, IMS-DB/DC, IDMS, JCL/PROC, REXX, CLIST.