Spark Developer Resume
3.00/5 (Submit Your Rating)
SUMMARY:
- CCAA 175 Certified Hadoop and Spark Developer, Oracle Certified Associate with 8+ years of IT experience in software development.
- Ability to design, develop, deploy and support solutions using Agile Scrum methodology that leverage the Client and good understanding of various phases in Software Development life cycle (SDLC).
- Hands on experience in working with Hadoop Ecosystems, Spark (Scala Programming), Hive, Flume, Kafka, Spark Streaming and Sqoop
- Expertise on building Spark/ Saprk SQL API for Data Acquisition from various partner systems and ETL processing
- Proficiency in handling File Formats (Sequence Files, ORC, Text/CSV, Avro, Parquet) and analyzed using HiveQL.
- Hands on experience in writing Ad - hoc Queries for moving data from HDFS to HIVE and analyzing the data using HIVE QL .
- Working experience in importing and exporting data using Sqoop from Relational Database Systems (RDBMS) to HDFS
- Genereate End user reports using Tableau
- Build modules with SBT and Maven and Continuous integration alongside with Jenkins, Project management on Agile Methodology, Issue tracking and Release management with JIRA, ITSM, Peregrine and HP ALM.
- Proficient in Release and Change Management process (ITSM) and Peregrine Incident Management Tools.
- Knowledge on Basel III and CCAR capital requirement regulatory framework on bank capital adequacy, stress testing.
- Implementation of Mutual funds, Derivatives, Exception Management business logic in TLM (Transaction Life Cycle Management) reconciliation Tool. Expertise over Schema Design and KB file Configuration as well as Recon Admin and Smart Studio, Configuration Transfer (CT).TLM AAS, BMC Control-M (Job Scheduler), PL/SQL, UNIX Shell Scripting, BMC Control-M and Autosys Job Scheduling utilities.
- Coordinate with project and deployment team to avoid risk and impact over reliant applications that may popup during code deployment process.
- Preparation of BRD, FSD (high level design documents) and interacting with Business Associates, prepare TDD (low level design documents) document to ensure the Product quality as per the business requirements.
- Receptive, self-motivated professional providing creative solutions, working constructively as a team member or independently.
- Team Lead with good interpersonal & management skills.
TECHNICAL SKILLS:
Big Data: Spark, Scala, HDFS, YARN,HBase, Hive, Tableau, Oozie,Kafka, Flume, Sqoop, K-SQL,Putty
Database: Oracle, Sybase andMysql
Scripting: Unix Shell Scripting, PL-SQL
Tools and Utilities: SBT, Jira, Maven, HP ALM, ITSM, TLM 2.7V Components, BMC Control - M, CAAutosys R1, SSMA, SQL Developer 3.2v, HP QC10V, SQL WAYS.
PROFESSIONAL EXPERIENCE:
Confidential
Spark Developer
Roles and Responsibilities
- Interaction with business Stake holders to gather requirements and prepare functional and technical design documents and produce for approval
- Migrate the Finance Modernization data from Legacy mainframe Applications to HDFS via spark ETL processing.
- Categorize in scope CDE to corresponding Key Value Groups specific to their functionality.
- Build a spark application to automate all the Critical data element for the application respectively and write to HDFS
- Data in mainframe VSAM files is first moved to Data Lake with Informatica BDE Version. Then categorized to rows in HDFS master data storage as key value groups.
- All the CDE are casted as per the datatype specified in Data dictionary and populated in column level in domain view
- Implemented various performance tuning techniques such as joins with same practitioner to avoid shuffle, Broadcast dataframes etc.
- Built De-orchestration scripts to automate entire process
- Automate Data quality checks for all the implemented CDE via spark applications
- Hands-on experience in using Hive partitioning, bucketing and execute different types of joins on Hive tables and implementing Hive SerDes like JSON and Avro.
- Log JIRA for issues tracking and release management
- Continuous Integration using Maven to repository though Jenkins Build and release process. schedule workflows jobs via Autosys
Confidential
Spark Developer
Roles and Responsibilities
- Interaction with business Stake holders to gather requirements and prepare functional and technical design documents and produce for approval
- Build Spark for Data Acquisition in structured, unstructured and compression codecs (Snappy,.gz..), ETL Processing such as masking sensitive Obligor and Facility corevitals.
- Data Quality Authentication such as general ledger accounting keys, associated collaterals, Reportingunit, Department Id based on origin. Reporting data quality.
- Develop Sqoop Scripts to extract data from RDBMS onto HDFS for monthly and Incremental loading in Daily process
- Utilizing oozie to schedule workflows.
- Creation of subscription in Hive for End user reporting
- Triage Analysis and Publish overall risk Metrics with stake holders on monthly basis.
- Worked on both External and Managed HIVE tables for optimized Performance
- Hands-on experience in using Hive partitioning, bucketing and execute different types of joins on Hive tables and implementing Hive SerDes like JSON and Avro.
- Process stock transactions based on facility and Obligorusing Kafka to Phoenix via spark streaming.
- Log JIRA for issues tracking and release management
- Continuous Integration using Maven to repository though Jenkins Build and release process.
Confidential
Hadoop Developer
Roles and Responsibilities
- Primary responsibilities include building scalable distributed data solutions using Hadoopecosystem
- Experienced on loading and transforming of large sets of structured, semi structured and unstructured data. Imported data using Sqoop to load data from MySQL to HDFS on regular basis.
- Develop and Automate Sqoop jobs on YARN and Hadoop clusters to produce daily and monthly reports as per business requirements
- Worked on both External and Managed HIVE tables, Implemented Partitioning, Dynamic Partition, Buckets for optimized performance.
- Hands-on experience in using Hive partitioning, bucketing and execute different types of joins on Hive tables and implementing Hive SerDes like JSON and Avro.
- Developed HIVE scripts for analyst requirements for analysis.
- Working close together with QA and Operations teams to understand, design, and develop and end-to-end data flow requirements.
- Utilizing oozie to schedule workflows.
- Handled importing of data from various data sources, performed transformations using Hive, loaded data into HDFS and Extracted the data from MySQL into HDFS using Sqoop
Confidential
TLM L3 Production Support)
Roles and Responsibility
- Involved in L3 Support for TLM Reconciliations and Feed Processing Jobs Scheduled in BMC Control-M.
- Monitor progress against established schedules and resolving deviations.
- Involved in business logic configuration using TLM software, System Testing for the new modules/reconciliations, enhancements, and maintenance of already developed modules
- Involved in Build & Unit Testing of TLM Configuration Enhancements and Jobs Scheduled in BMC Control-M.
- Involved in developing the UNIX shell scripting and database scripting to process the records from the backend.
- Creation of Change request for all the code deployments in ITSMand incidents through peregrine to monitor all deployment.
- Involved in Weekly transaction purging process.
- Involved in Enhancements of already existing backend scripts.
- Report project status to the JPMC Project Manager on a weekly basis.
- Reporting project status to Vendor Management and escalating issues, as appropriate.
- Involved in coordination of Code deployments and performance testing.
Confidential
TLM Configurer
Roles and Responsibility
- Involved in requirement analysis, client interaction for gathering the requirements and confirming the business logic design
- Produce/prepare TDD on various reconciliations. It includes understanding and designing of match logics, Customized BLT’s,dashboards, schema layout, defining attributes and data type and more.
- Developed UNIX and T-SQL scripts to initiate STP Workflow Manually, Automation of Feed loading in TLM.
- Configuration ofSchema, KB, STLFetch and matching logic, Schema and Dashboard Design for Cash, Position and Balance Reconciliations.
- Preparation of Unit test plans for the build and testing the same.
- Developed UNIX and SQL scripts to initiate STP Workflow Manually, Automation of Feed loading in TLM.
- Part of Unit, System Integration testing and UAT.
- Knowledge Transfer to Stake Holders and SIT team.