Hadoop | Spark Engineer Resume
SUMMARY
- 13+ Years of experience in Analysis, Design, Development, testing and Implementations of Bigdata applications and software applications.
- 5+years of experience in Hadoop, Spark, Java technologies and 8+ years of experience in Mainframe, SQL Technologies. Solid technical working experience as well as domain knowledge in Banking and Financial, Credit Card and Insurance domains.
- Worked on Machine learning POC's
- Predict Digital Wallet transaction approvals and declines near real time to achieve advance alerting mechanism.
- Predict and recommend the CARE Enrollment using the historical ageing information.
- Experience in all stages of SDLC (Agile, Waterfall), writing System Requirement Specification and design principles.
- Have very good technical, problem solving and communication and interpersonal skills and can manage and working with team.
PROFESSIONAL EXPERIENCE
HADOOP | SPARK ENGINEER
Confidential
Responsibilities:
- Developed Hive queries, Spark Data frames and RDD for different Data aggregation queries
- Developed Spark process using Datasets, Data frames, JavaRDD and JavaPairRDD for reading & writing HDFS files and Hive tables.
- Implemented Business logic in Spark using mapToPair, flatMap methods in Spark.
- Delivered Spark Process for Comparing two Hive tables and writing the delta records to HDFS files.
- Implemented Spark Process for transposing Rows to Columns and vice versa.
- Built Continuous Integration and Development (CI/CD) pipeline for Lucy
- Performed different types of Transformations and actions on the RDD to meet the business requirements
- Wrote Hive queries for data analysis to meet the business requirements.
- Wrote Map reduce programs for formatting and storing the final score data into HDFS
- Developed Shell scripts for consolidating final ouputs and SFTP into mainframe servers
- Worked on Solace and Spring boot interaction to consume new POT elements and update into Risk DB
- Write Shell scripts to automate application deployments and configurations monitoring
- Palladium (pre - production environment), Silver & Gold set ups (E1 & E2) for ProbC use case.
- Involved in creating Hive tables and loading them with data and writing Hive queries.
- Provide design recommendations and thought leadership to sponsors /stakeholders that improved review processes and resolved technical problems.
- Partnering with DevOps Infrastructure Teams and Creating Application Release and Development Plans for complex Deviant, Systems, Pre-Production and Production. (Removed unwanted processes, overheads, improved rate of deliverable by addressing risk early phase of planning)
Environment: Mapr,EnterPrise Cloud Platform, Spark SQL, PySpark, Hive, Shell Script, Map Reduce, Java, Solace, Spring Boot, Event Engine, Eclipse, IntelliJ, Github, Maven, Jenkin
Confidential
HADOOP | SPARK ENGINEERResponsibilities:
- Implemented various Hive queries for extracting the data from different Hive tables.
- Designed and developed Map Reduce process for processing the data from different input tables and applying the business logic before storing data.
- Created HBase tables to store unstructured data for online availability.
Designed and developed Map Reduce process for Load & Delete data into HBase tables
- Implemented various Oozie workflows to process both Hive queries extraction and Map Reduce process.
- Developed various shell scripts to perform pre and post validations for whole process.
- Provide testing support and resolution.
Environment: MapR, Hive, Hbase, Shell Script, Map Reduce, Oozie, Java, Eclipse, IntelliJ, Github, Maven, Jenkin
BIG DATA DEVELOPER
Confidential
Responsibilities:
- Work closely with the business and analytics team in gathering the system requirements.
- Involved in high-level architecture design of tables and databases
- Developed various Hive Queries to process data from various source systems tables.
- Created HBase tables to store unstructured data for reporting purposes.
- Designed and developed Map Reduce process for Loading the data into HBase tables
- Developed various shell scripts to perform pre and post validations for whole process.
- Provide testing support and resolution to validate the final results between DB2 and HBase.
Environment: Mapr, Hive, Map reduce, HBase, Shell Script, Sqoop, Oozie, Java, Eclipse, InteliJ, GitHub, Maven, Jenkin
BIG DATA DEVELOPER
Confidential
Responsibilities:
- Gathering data requirements and identifying sources for acquisition.
- Migrated complex Mainframe process to Big data world
- Worked on data ingestion into Big data platform
- Involved in high-level architecture design of tables and databases for enterprise future use.
- Worked on Kafka and Spring boot interaction to consume various demographic and financial elements and update into Risk DB
- Developed various Hive Queries to process data from various source systems tables.
- Developed Map Reduce Process to process Multiple tables and writes data into 360 degree view of customer.
- Created Oozie workflows for execution of multiple jobs in production with predefined SLA times.
- Developed Unix scripts for corn jobs in production.
- Created Hive tables to store structured data for downstream processing.
- Developed process for updating/inserting the Hive Data on daily basis.
- Provide testing support and resolution.
- Designed and developed testing tool for validations.
Environment: Hadoop, HDFS, MapReduce, Oozie,Hive, Shell Scripting, Core Java, Kafka and Spring Boot
JAVA DEVELOPER
Confidential
Responsibilities:
- Identified and Migrate various process and services in existing Mainframe applications, which can migrate to Java
- Preparing Design documents for the flow of changes based on the analysis as per the quality standards.
- Coded and unit tested application enhancements.
- Re-engineering and Performance tuning of high CPU utilization modules in Mainframes.
- Developed Web service for automating the test process using Java and rest API’s
- Developed the enhancements for Soap API services.
- Performed regression and system-level testing to verify software quality and function before it was released.
- Supporting the Implementation related activities
Environment: Core Java, JSP,HTML, Soap and Rest API, IMS connect, IBM S/390, MVS, JCL, COBOL, SQL, DB2,IMS DB/DC, EXPEDITOR, SORT, EZY, CHGMAN
MAINFRAME | SQL TECH LEAD
Confidential
Responsibilities:
- Strategically architect, design, develop, and implement efficient information systems and/or operations systems in support of core enterprise functions.
- Interacting with the clients to gain organizational commitment for all systems and software plans, as well as evaluate and guide the selection of technologies required to complete those plans.
- Analyzing the system to identify the opportunities to optimize the system.
- Migrated various processes from Mainframe to Java.
- Identified various process and services in existing Mainframe applications, which can migrate to Java using IMS connect
- Identifying the process that can be automated and thereby reducing manual effort.
- Coordinate with offshore development team(s) and different groups of interfacing system to identify priorities and update scope and delivery schedule
- Preparing Design documents for the flow of changes based on the analysis as per the
- Analysis on the specifications provided by the client.
- Coding the enhancements.
- Doing peer reviews for the design documents and the programs.
- Developing test plans and Testing of enhancements.
- Documentation as per the quality procedures.
- Supporting SIT, UAT & Regression.
- Taking care of the implementation activities
Environment: IBM 3090,MVS, COBOL, JCL, SQL, DB2, IMS-DB,, Change man, Xpeditor, File-Aid, Spufi, QMF, DB2 BMC utility, Jobtrac, Infoman, MQ Series, Java
