Hadoop/spark Developer Resume
Richmond, VA
SUMMARY:
- Over 11 years of experience, including 3 years on Big Data Hadoop Eco Systems with hands - on project experience in health care domain.
- Cloudera certified Spark and Hadoop developer
- Extensive experience on SQL&PL/SQL, Hadoop and Linux Shell Scripting
- Expertise in developing Spark applications using Scala
- Exposure to data quality analysis, data migration and data validation.
- Migrating the coding from Hive to Apache Spark and Scala using SparkSQL, RDD
- Working experience on Spark, Spark SQL, RDD, Data Frames and Datasets
- Hands on experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems and vice-versa.
- Good Experience in file handling operations on Unix.
- Professionally adaptable to 'get on board' quickly with new technologies and new assignments, able to work independently or in a group/team setting to meet deadlines while handling multiple assignments.
- Extensively involved in Requirement analysis, Code development and Bug fixing.
- Expertise in Hive Query Language and debugging Hive issues.
- Expertise in Sqoop and Flume
- Extremely motivated with good inter-personal Skills have ability to work in strict deadlines
- Accomplished facilitator in understanding
- Proficiency in SDLC Life Cycle, Agile and Waterfall models
- Coordinating with offshore team on technical requirements, design, test plan, test results, pre-implementation activities and deployment
TECHNICAL SKILLS:
Big data ecosystem: Hadoop, MapReduce, HDFS, Hive, Flume, Kafka, Sqoop, Hbase, Zookeeper, Oozie, Apache Spark, SparkSQL, Spark Streaming, RDD, Data Frames, JSON, AVRO
Programming/scripting languages: Scala, Linux shell scripts, Cobol, JCL, XML
Databases: IBM-DB2, MySQL, IDMS, IMS-DB
Operating systems: Unix, centos 6.7, Z/OS, MVS, Windows …
Development tools: SUBLIME 2, Eclipse, AVRO-TOOLS
Collaboration Tools: JIRA, Rational Team Concert, Rational ClearQuest, SharePoint
PROFESSIONAL EXPERIENCE:
Hadoop/Spark Developer
Confidential, Richmond, VA
Responsibilities:
- Involved in developing roadmap for migration from legacy system to Hadoop cluster.
- Create, validate and maintain scripts to load data using Sqoop manually.
- Load and transform large sets of structured, semi structured and unstructured data coming from various downstream systems.
- Migrated data between RDBMS and HDFS/Hive with Sqoop.
- Create, validate and maintain scripts to extract and transform data from MySQL to flat files and JSON format.
- Create Oozie workflows and coordinators to automate Sqoop jobs weekly and monthly.
- Worked on reading multiple data formats on HDFS using Apache Spark.
- Wrote Scala scripts for spark to perform operations like data inspection, cleaning, load and transforms the large sets of structured and semi-structured imported data
- Involving in Migrating the coding from Hive to Apache Spark and Scala using SparkSQL, RDD
- Developed Spark with Scala and Spark sql for testing and processing of data
- Developed, validate and maintain HiveQL queries.
- Designed Hive tables to load data to and from external HDFS datasets
- Hands on using Partitions, Bucketing concepts in Hive and designed both Managed and External tables in Hive for optimized performance.
- Managing and scheduling jobs on a Hadoop cluster.
Environment: Hadoop, HDFS, Apache Spark 1.6,SparkSQL, Unix, Hive, Sqoop, Flume, Scala, Oozie, DB2 SQL, kafka
Senior System Analyst
Confidential, Albany, NY
Responsibilities:
- As Senior System Analyst, managed and coordinated multiple systems/teams involved to ensure successful delivery of the efforts as per the Business Requirements
- Ensure on-time delivery for the SDLC phases for projects and SSCRs (Small System Change Requests)
- Perform weekly work forecast, LOE reviews and related negotiations
- Represent the CHIPS VA provider and CS90 NY provider teams in the weekly Client Portfolio and Stake Management calls
- Identify and raise Risks, Issues and dependencies to the stakeholders and come up with mitigations plans in a timely manner
- Review the deployment plan and document implementation strategies with the Confidential management and SME's. Discuss risks, issues and dependencies in the implementation procedures in a deployment across applications
- Prepare and review Level of Estimate in hours for each of the efforts executed at Confidential . Based on industry standards the LOE is prepared with clear breakup for each phase of the project and sometimes at activity level
- Arrange internal JAD sessions and come up with the best possible design approach to the many projects and SSCRs
- Interacting with SME's, Business, Application owners/end users to determine application requirements, user problems, participating in architecture and design activities
- SPOC for the CHIPS VA provider and CS90 NY provider team for the release management activities for the Confidential CHIPS Virginia System and CS90 Newyork systems.
- Owner of the Procedures and Process documentation for the Confidential account and its maintenance
Environment: Z/OS, MVS, COBOL, JCL/PROC, VSAM, Cobol-XML, DB2 SQL, SPUFI, IMS, REXX, CLIST, VB scripting, ClearQuest, Share-point
Batch Process Lead & Application developer
Confidential
Responsibilities:
- Studying the different systems that exist in Health-care insurance and their functional impact.
- Understanding the client requirements and preparation of technical documents.
- Working and implementing the code in the mainframe environment.
- Understanding the interfaces for claims software and PLASM language.
- Maintenance, enhancement and developments of programs coded in COBOL as per the requirements.
- Working on the Incidents primarily concerning claims (maintenance support system).
- Supporting the high-priority live batch flow system.
Environment: Z/OS, MVS, COBOL, IMS-DB/DC, IDMS, JCL/PROC, REXX, CLIST.