We provide IT Staff Augmentation Services!

Hadoop | Spark Engineer Resume

2.00/5 (Submit Your Rating)

SUMMARY

  • 13+ Years of experience in Analysis, Design, Development, testing and Implementations of Bigdata applications and software applications.
  • 5+years of experience in Hadoop, Spark, Java technologies and 8+ years of experience in Mainframe, SQL Technologies. Solid technical working experience as well as domain knowledge in Banking and Financial, Credit Card and Insurance domains.
  • Worked on Machine learning POC's
  • Predict Digital Wallet transaction approvals and declines near real time to achieve advance alerting mechanism.
  • Predict and recommend the CARE Enrollment using the historical ageing information.
  • Experience in all stages of SDLC (Agile, Waterfall), writing System Requirement Specification and design principles.
  • Have very good technical, problem solving and communication and interpersonal skills and can manage and working with team.

PROFESSIONAL EXPERIENCE

HADOOP | SPARK ENGINEER

Confidential

Responsibilities:

  • Developed Hive queries, Spark Data frames and RDD for different Data aggregation queries
  • Developed Spark process using Datasets, Data frames, JavaRDD and JavaPairRDD for reading & writing HDFS files and Hive tables.
  • Implemented Business logic in Spark using mapToPair, flatMap methods in Spark.
  • Delivered Spark Process for Comparing two Hive tables and writing the delta records to HDFS files.
  • Implemented Spark Process for transposing Rows to Columns and vice versa.
  • Built Continuous Integration and Development (CI/CD) pipeline for Lucy
  • Performed different types of Transformations and actions on the RDD to meet the business requirements
  • Wrote Hive queries for data analysis to meet the business requirements.
  • Wrote Map reduce programs for formatting and storing the final score data into HDFS
  • Developed Shell scripts for consolidating final ouputs and SFTP into mainframe servers
  • Worked on Solace and Spring boot interaction to consume new POT elements and update into Risk DB
  • Write Shell scripts to automate application deployments and configurations monitoring
  • Palladium (pre - production environment), Silver & Gold set ups (E1 & E2) for ProbC use case.
  • Involved in creating Hive tables and loading them with data and writing Hive queries.
  • Provide design recommendations and thought leadership to sponsors /stakeholders that improved review processes and resolved technical problems.
  • Partnering with DevOps Infrastructure Teams and Creating Application Release and Development Plans for complex Deviant, Systems, Pre-Production and Production. (Removed unwanted processes, overheads, improved rate of deliverable by addressing risk early phase of planning)

Environment: Mapr,EnterPrise Cloud Platform, Spark SQL, PySpark, Hive, Shell Script, Map Reduce, Java, Solace, Spring Boot, Event Engine, Eclipse, IntelliJ, Github, Maven, Jenkin

Confidential

HADOOP | SPARK ENGINEER

Responsibilities:

  • Implemented various Hive queries for extracting the data from different Hive tables.
  • Designed and developed Map Reduce process for processing the data from different input tables and applying the business logic before storing data.
  • Created HBase tables to store unstructured data for online availability.

    Designed and developed Map Reduce process for Load & Delete data into HBase tables

  • Implemented various Oozie workflows to process both Hive queries extraction and Map Reduce process.
  • Developed various shell scripts to perform pre and post validations for whole process.
  • Provide testing support and resolution.

Environment: MapR, Hive, Hbase, Shell Script, Map Reduce, Oozie, Java, Eclipse, IntelliJ, Github, Maven, Jenkin

BIG DATA DEVELOPER

Confidential

Responsibilities:

  • Work closely with the business and analytics team in gathering the system requirements.
  • Involved in high-level architecture design of tables and databases
  • Developed various Hive Queries to process data from various source systems tables.
  • Created HBase tables to store unstructured data for reporting purposes.
  • Designed and developed Map Reduce process for Loading the data into HBase tables
  • Developed various shell scripts to perform pre and post validations for whole process.
  • Provide testing support and resolution to validate the final results between DB2 and HBase.

Environment: Mapr, Hive, Map reduce, HBase, Shell Script, Sqoop, Oozie, Java, Eclipse, InteliJ, GitHub, Maven, Jenkin

BIG DATA DEVELOPER

Confidential

Responsibilities:

  • Gathering data requirements and identifying sources for acquisition.
  • Migrated complex Mainframe process to Big data world
  • Worked on data ingestion into Big data platform
  • Involved in high-level architecture design of tables and databases for enterprise future use.
  • Worked on Kafka and Spring boot interaction to consume various demographic and financial elements and update into Risk DB
  • Developed various Hive Queries to process data from various source systems tables.
  • Developed Map Reduce Process to process Multiple tables and writes data into 360 degree view of customer.
  • Created Oozie workflows for execution of multiple jobs in production with predefined SLA times.
  • Developed Unix scripts for corn jobs in production.
  • Created Hive tables to store structured data for downstream processing.
  • Developed process for updating/inserting the Hive Data on daily basis.
  • Provide testing support and resolution.
  • Designed and developed testing tool for validations.

Environment: Hadoop, HDFS, MapReduce, Oozie,Hive, Shell Scripting, Core Java, Kafka and Spring Boot

JAVA DEVELOPER

Confidential

Responsibilities:

  • Identified and Migrate various process and services in existing Mainframe applications, which can migrate to Java
  • Preparing Design documents for the flow of changes based on the analysis as per the quality standards.
  • Coded and unit tested application enhancements.
  • Re-engineering and Performance tuning of high CPU utilization modules in Mainframes.
  • Developed Web service for automating the test process using Java and rest API’s
  • Developed the enhancements for Soap API services.
  • Performed regression and system-level testing to verify software quality and function before it was released.
  • Supporting the Implementation related activities

Environment: Core Java, JSP,HTML, Soap and Rest API, IMS connect, IBM S/390, MVS, JCL, COBOL, SQL, DB2,IMS DB/DC, EXPEDITOR, SORT, EZY, CHGMAN

MAINFRAME | SQL TECH LEAD

Confidential

Responsibilities:

  • Strategically architect, design, develop, and implement efficient information systems and/or operations systems in support of core enterprise functions.
  • Interacting with the clients to gain organizational commitment for all systems and software plans, as well as evaluate and guide the selection of technologies required to complete those plans.
  • Analyzing the system to identify the opportunities to optimize the system.
  • Migrated various processes from Mainframe to Java.
  • Identified various process and services in existing Mainframe applications, which can migrate to Java using IMS connect
  • Identifying the process that can be automated and thereby reducing manual effort.
  • Coordinate with offshore development team(s) and different groups of interfacing system to identify priorities and update scope and delivery schedule
  • Preparing Design documents for the flow of changes based on the analysis as per the
  • Analysis on the specifications provided by the client.
  • Coding the enhancements.
  • Doing peer reviews for the design documents and the programs.
  • Developing test plans and Testing of enhancements.
  • Documentation as per the quality procedures.
  • Supporting SIT, UAT & Regression.
  • Taking care of the implementation activities

Environment: IBM 3090,MVS, COBOL, JCL, SQL, DB2, IMS-DB,, Change man, Xpeditor, File-Aid, Spufi, QMF, DB2 BMC utility, Jobtrac, Infoman, MQ Series, Java

We'd love your feedback!