Senior Hadoop Developer Resume
Malvern, PA
SUMMARY
- 10+ years of professional IT experience which includes experience in Hadoop/Bigdata ecosystems and related Technologies.
- Currently working as a Senior developer/Technical Lead in of Hadoop/Bigdata analytics using the Cloudera Distribution and Amazon Web Service EC2 platforms.
- Worked as a Technical Lead/Senior Developer/Onsite coordinator in the area of Mainframe/Java/ETL Technologies with various clients and domain.
- Excellent Experience in Hadoop architecture and various components such as HDFS Job Tracker Task Tracker NameNode Data Node and MapReduce programming paradigm
- Expert in the developing Spark program using program interfaces Scala and Python (pyspark)
- Extensive experience in ETL process, data architecture including data ingestion pipeline design, Hadoop information architecture, data modeling and data mining and advanced data processing. Experienced in optimizing ETL workflows
- Proficient in designing and coding the Oozie workflows in the ETL and Conformance layer in Hadoop ecosystem
- Expert in Hive, Impala and HiveUDFs
- Exported and imported large volume of data from Hadoop ecosystem to various relational databases using Sqoop
- Very good understand on AWS EMR/EC2, AWS S3, CloudFormation, CloudWatch, AWS lambda function etc
- Proficient in Atlassian tools like Bamboo, Bitbucket, Github and JIRA
- Very good understanding on Distributed computing, Linux distributions and Mainframe systems.
- Expertise in IBM Mainframe with deep knowledge in mainframe based applications and mainframe tools
- Expertise in troubleshooting and proficient to lead the team to fix large production issues
- Proficient in Project Management, Production support, Application Development, Programming, System Analysis, Software Quality assurance and change management process with various clients
- Conversant with all phases of project Life Cycle including Requirement gathering, Analysis, Design, Development, Testing, Implementation, Software quality standards and configuration management and change management and Quality procedures.
- Expertise in Application development projects and very good exposure in the development methodologies like Agile, Water fall etc.
- Expertise in handling support and maintenance project also hands on experience in the ticket tracking tools like HP SMPO, ITSM and Remedy, JIRA.
- Hands on experience in the migration of mainframe application to other technologies like SAP, UNIX etc and Re - hosing and decommissioning mainframe to Micro focus enterprise server.
- Handled small/medium size team, good at operational planning, work allocation, tracking and reviewing
- Have good combination of technical, communication and interpersonal skills, which provide the ability to be an effective mediator between programmers, end-users and clients.
- Highly motivated with the ability to work effectively in teams as well as independently.
- Experienced in executingOnsite-Offshoreproject execution models, Proficient in document management, able to prioritize and complete multiple tasks.
- Have intense knowledge in various domains like retail, banking, and insurance, healthcare.
TECHNICAL SKILLS
Platforms/frameworks: Hadoop 2.7,IBM S/390, IBM PC Compatibles
Operating systems: Linux,OS/390,Windows 10/7/XP/2000/Server,MS-DOS
API: Spark 1.6,2.x, MapReduce
Programming Language: Python, Scala, Java, VS COBOL, JCL, Easytrive,SAS
Scripting Language: Korn shell/Unix shell scripting,XML,SQL
Workflow: Oozie
Databases: Hive,Impala,DB2,Oracle,IMS DB
ETL Tool: Sqoop, Flume,Kafka
Web Interface: Hue
File systems: Avro files, Parquet files, HDFS,VSAM
OLTP: CICS, IMS DC/TM
Middleware: MQ Series
Tools: /Technologies: Spring Tool Suite,Eclipse,Crucible,ChangemanEndavor,Panvalet, Panvalet, Xpeditor,DB2/VSAMFileaid, Paltinum Startool, SAR,Jobtrac, SPUFI,QMFTape Management system (TMS), OPC schedulerAbendaid, DADS, IBM debugger,Mainframe Express
Tracking Tools: AtlassianTools - BitBucket,Bamboo,JIRA
PROFESSIONAL EXPERIENCE
Confidential, Malvern PA
Senior Hadoop Developer
Responsibilities:
- Migration of the various client score models developed in On-Prem Hadoop to AWS EMR
- Refactor the existing score model logic from Warehouse tables to Enterprise table and map the Warehouse logic to Enterprise logic
- Code the new score model programs in Pyspark and Spark-Scala
- Covert the existing Pyspark and Spark Scala programs from Spark 1.6 to Spark 2.2 version.
- Migrate the existing Sqoop tables from On-Prem to S3 bucket and build the new Sqoop pipeline for the newly added enterprise tables to S3 bucket.
- Convert the existing integration suits running in Impala to Hive/S3 bucket.
- Create and customize the CloudFormation Templates using Troposphere and spin up the AWS EMR cluster
- Integrate, build and deploy the CloudFormation/Delete templates like S3-Copy,Create Stack and Delete Stack using Bamboo.
- Create the CloudWatch events for the AWS EMR logs and integrate the CloudWatch logs to the Splunk
Confidential, Pennigton NJ
Senior Hadoop Developer
Responsibilities:
- Design the DataLake to pull the HMDA Loan details of various client from the upstream system like Peaks, nCino
- Design and implement the sqoop process to pull the client data from various Oracle data bases to Hadoop environment of BOFA.
- Implement ETL process and conformance code for the HMDA Data Lake
- Design and Implement the Oozie workflow to import and export client’s loan information to various loan processing and data analytical systems in BOFA
- Create and Hive tables in the Hadoop data hub region and store the sqoop data in the parquet format.
- Design and code the conformance logic using Spark-Scala which can be used for target or consuming systems.
- Optimize the Spark-Scala, Spark-SQL code in the conformance layer for the process improvement.
- Implement the Oozie coordinator and schedule the daily/weekly/monthly jobs.
- Create the test suits using JUnit and perform the unit, integration and end to end testing using the JUnit in QA and SIT regions.
- Optimizing the Hive queries using Partitioning and Bucketing techniques, for controlling the data distribution.
Confidential, Malvern, PA
Technology specialist
Responsibilities:
- Design the ETL process to bring the client score details from Teleaf, Data warehouse and enterprise system to Vanguard Hadoop ecosystem.
- Discuss with business users to understand and clarify the business requirement and prepare the design documents
- Design, code and implement the sqoop process and import the score details Hadoop data hub.
- Perform cleanse and validation of the imported data convert to the Avro file format which are accessible to the Hadoop data mart environment.
- Make the necessary changes to cleanse and validate programs using spark-scala.
- Design and code the score calculation logic for the Vanguard client in using pyspark and execute the pyspark programs in Hadoop data mart environment.
- Design and implement the Oozie workflow for the daily/weekly/monthly client score calculation and web interaction reports.
- Implement the Oozie coordinator and schedule the daily/weekly/monthly jobs.
- Create the test suits using pyspark and perform the unit, integration and end to end testing using the pyunits.
- Convert the Avro files in the Hadoop data hub to parquet format using the Hive scripts.
- Import the data from Oracle and DB2 database to Hadoop ecosystem using sqoop.
- Create the Hive tables in Hadoop data mart environment and validate the performance of Hive and Impala queries against the master tables.
- Optimizing the Hive queries using Partitioning and Bucketing techniques, for controlling the data distribution.
- Fine tune the pyspark codes for the optimized utilization of Hadoop resources for the production run.
- Execute the comparison test in the production region and fine tune the end results to ensure the accuracy.
- Trouble shoot and fix the daily Oozie workflow failure and implement the permanent fix.
- Analyze the Java-MapReduce program and prepare the analysis documents and perform the feasibility study to convert the Java-MapReduce program to Spark-Python(pyspark).
- Prepare the High level/low level design document for the conversion Java-MapReduce code to pyspark.
- Code the programs in pyspark those are in Java and perform the Unit/Integration/Regression and comparison testing to ensure that the newly converted code have the same functionality and performance with the code in Java.
- Mentor the juniors and provide the application training for the new joiners in the team.
Confidential
Project Lead
Responsibilities:
- Go thru the various BRD's and understand the business requirement also discuss with client to get clarification on the requirement.
- Prepare the Technical Spec and send to client for approvals.
- Provide the estimation for Design, coding, testing and Implementation and support phases for each releases.
- Design and code the DB2 stored procedures to execute from Java Interface.
- Allocate the task to the team members and review their deliverables.
- Coordinate the analysis/coding/testing activities with team to ensure the traceability to requirements is maintained.
- Review the code, test results and prepare the review and defect logs.
- Attend the daily/weekly status call with onsite/client and provide the updates and answer the queries.
- Prepare the daily, weekly and monthly status report also attend the SMR and prepare the monthly SMR MOM.
- Responsible for providing and coordinating the cross functional for the peers.
- Identify value add enhancement for the application, create the impact analysis document and submit to the client for approval.
- Present the RCA reports to the client and suggest the enhancement of the application module to fine tune the application performance.
- Analyzing the system performance.
- Review the quality documents prepared by the team as per the change management process.
Confidential, New Jersey
Project Lead
Responsibilities:
- In the role of Project Lead, responsible for generating the daily, weekly and monthly status report for the L2 and L3 tickets.
- Resolving L3 tickets for the UBS’s Banking application.
- Work on major and minor enhancement also provides the estimation to the client.
- Identify value add enhancement for the application, create the impact analysis document and submit to the client for approval.
- Allocate the task to the team members and review their deliverables.
- Initiate and coordinate the bridge call for Critical and High priority incidents.
- Provide 24/7 on-call support and responsible as the 1st level escalation manager for the on-call support
- Responsible for tracking of trouble tickets and resolution of escalated support incidents
- Answers technical support queries and implements fixes for application problems
- Responsible for providing and coordinating the cross functional for the peers.
- Coordinate the analysis/coding/testing activities with team to ensure the traceability to requirements is maintained
- Responsible for perform the root cause analysis for the recurring issues and provide the permanent fix.
- Present the RCA reports to the client and suggest the enhancement of the application module to fine tune the application performance.
- Attend the client meeting.
- Analyzing the system performance.
- Monitoring the Batch Cycles and flow of System jobs.
- Review the quality documents prepared by the team as per the change management process.