Senior Big Data Developer Resume
4.00/5 (Submit Your Rating)
New, JerseY
SUMMARY
- 13+ Years of Development and Design experience in Data warehousing, ETL, Big Data using Hadoop echo systems Hive, Pig, Sqoop and SPARK, Python (PySpark) programming
- Experience in developing ETL programs in Python to extend the HIVE functionality
- Strong experience with Spark batch, and Spark streaming
- ETL experience in Hive
- Development experience with Scala for the batch processing and Spark streaming data
- Developed Spark RDD, Data Frames and SQL scripts for ETL transformations.
- Big Data ETL Design Solution and Development experience for Massive Data Warehousing HR Systems with 500+ Interfaces along with ARIBA, ADP Payroll interfaces
- Area of expertise is enabling scalable ETL data pipelines using both Hadoop Technologies(MR/Pig/Hive)
- Solid experience in the Hadoop and HDFS systems
- Experience in major distribution like Cloudera(CDH) and Hortonworks Distribution (HDP)
- Good understanding of Apache HBase
- Experience of Importing data using sqoop into HDFS from various Relational Database Systems
- Hands on experience in implementing business logic and optimizing the query using Hive SQL
- Experience with Oozie Workflow Engine in running workflow jobs with actions that run Hadoop Pig jobs
- Data ingestion into HDFS using the Hadoop File systems commands
- Very Good Knowledge in Dimension Data Modeling
- Data warehousing and ETL development experience in Oracle SQL Informatica, SSIS and Unix
- Extensive experience in requirement gathering, requirement design, technical design docs
- Extensive Experience in Informtaica ETL development project in a healthcare domain
- Extensive experience in MS SQL, Oracle SQL, Database programming (T - SQL, PL/SQL)
- Worked as MS SQL Optimization & Performance Tuning expert for SQL Server applications
- More than 8 + Years’ experience in Banking, Health Care, Pharmaceuticals, Automotive domain
- 10 + Years of experience all aspects of data applications - Data Migration, and Data Modeling
- Development experience ETL solutions which involved creating debugging and executing Informatica Mappings, Sessions, Tasks and Workflow - Informatica PowerCenter 9.6
- Extensive experience in Unix Shell Scripting to automate the batch processing
- Business analytics experience on Cognos and Business objects
- Extensive experience in Tuning the Complex ETL Mappings and SSIS Packages
- Extensive experience in Data warehouse Architect life cycle
- Extensive experience as ETL Architect - Created Complex ETL Framework Solutions
- Extensive experience in Dimension Modeling including Slowly Changing Dimensional modelling(SCD)
- Extensive experience in Star Schema and Snowflake Schema
- Provided Informatica Training and setting up lab for Informatica training
- Proficient in analyzing and translating business requirements to technical requirements
- Oracle Certified Associate(OCA) - Oracle
- Certification in Sun Solaris Unix and Six Sigma
- Extensively working & worked with Agile and Waterfall model
- Good communication skills, interpersonal skills, self-motivated, quick learner, team player
TECHNICAL SKILLS
- Hadoop HDFS
- Spark
- Python
- Scala
- HBase
- Pig
- Hive
- Data Warehousing and ETL Architect Design Solution
- SSIS 2012
- Informatica Power Center 9.x/8.x/7.1
- Cognos 8
- Business Objects
- MS SQL Server 2008 and Oracle 10g/9i/11g SQL and PLSQL
- Dimensional Modeling
- Shell scripting
- Core Java
- Sun Solaris 9/8/7/2.6/2.5
- LINUX
- Windows XP Professional/2000/NT/98/95
- AUTOSYS
- UC4
- CONTROL-M
- TIDAL
- SVN
PROFESSIONAL EXPERIENCE
Confidential, New Jersey
Senior Big Data Developer
Responsibilities:
- Primary role is Senior Big Data Developer for HRIS Data warehousing systems
- Developed Warehousing programs for the new Workday Cloud to Data Warehousing System in HDFS environment
- Experience on Hadoop clusters using Cloudera distribution ( CDH)
- Scala programs development for batch processing
- Python scripting for ETL programs, validation and transformation
- Developed the solutions for Massive 500+ interfaces HRIS systems
- Created Requirement and Technical Design Documents for Big Data ETL Solution
- Mapping Functional requirements to Technical requirements.
- Loading and transforming large set of structured, semi structured and unstructured data
- Designed and Developed ETL programs using Python
- Written Complex Hive SQLs
- Managing data coming from 200+ different sources
- Loaded unstructured data into Hadoop File System( HDFS) using flume
- Written validation and data quality scripts
- Design and Development of new feeds to source data from new vendors and to deliver it to new clients.
- Analyzing data with Hive, Pig
- Python programs are developed to validate the files
- DesignedHivetables to load data to and from external files
- CreatingHivetables to import large data sets from various relational databases usingSqoopand export the analyzed data back for visualization and report generation by the BI team
Confidential, Jersey City, New Jersey
ETL Hadoop developer
Responsibilities:
- Creation and processing of HIVE External and Internal tables based on application requirements for Data Analytics
- Implementation of POC on Hadoop stack and different big data analytic tools, migration from different databases ( i.e. Teradata, Oracle, MySQL) to Hadoop.
- Work with OS Layer Security (User & group Policies, File permission, OS Authentication) according to Data compliance.
- Experience on Hadoop clusters using Hortonworks Distribution (HDP)
- Debug and solve issues with Hadoop as on-the-ground subject matter expert. This could include everything from patching components to post-mortem analysis of errors.
- Implementation of Cloudera cluster with High availability and standby solutions.
- Design and support of Data ingestion, Data Migration and Data processing for BI and Data Analytics.
- Responsible for developing data pipeline sqoop and pig to extract the data from weblogs and store in HDFS.
- Designed and implemented various metrics that can statistically signify the success of the experiment.
- Involved in developing Pig Scripts for change data capture and delta record processing between newly arrived data and already existing data in HDFS.
Confidential
Senior Performance Tuning Engineer
Responsibilities:
- Single point of contact for Across Confidential ETL and ERP SQL server database tuning
- Supporting DWH, ETL database environments
- SQL Server Performance application tuning export for Confidential DWH, ERP and ETL applications
- Working on deadlock resolution for Financial ERP product ( Sun INFOR10)
- Working on daily blocking issues
- Working on improvement on top poor SQLs -high elapsed time, high CPU time, high logical reads
- Working on interface slowness
- Working on proactive monitoring of databases
- Supporting across Confidential SQL Server applications ( Sun Infor10, Sun HFM, Bwise, Medgate, PANDA )
- Supporting Bwise application which is BO reporting application
Confidential
Informatica and SQL Developer
Responsibilities:
- Carry out functional, technical design reviews, as well as code reviews
- As an application architect, provide technical
- Implemented the many more automation for manual tasks.
- Assisting the team members for ETL rules and complex problems
- Preparing the technical notes for ETL
- Performing the role of Informatica administration.
- Preparing the POC for each design changes and enhancements
- Led the team of release 1 Lorenzo Data migration to Release 1.9 Lorenzo Data migration successfully
- Extensive experience developing Informatica mappings, mapplet, workflows, tasks, SQLs
Confidential
Oracle PL/SQL Developer
Responsibilities:
- Oracle/Unix technical analyst for five individual applications
- Gathered and defined user requirements.
- Created database triggers, procedures, functions using PL/SQL.
- Extensively written UNIX Shell scripts to extract the meta data from flat files to a specific format and import into Oracle database using SQL*Loader
- Customize or automate the applications to reduce the supporting hours using Unix Shell Scripts
- Resolving all Real Time Support(RTS) issues and provides production support activities
- Resolved the Visual Basic application problem like adding engine series, number and qualifier
- Columns into Embargo Screen which will block the data export to other system.
- Provided many solutions for Object Oriented Visual Basic application.
- Worked for a new interfaced which developed by SQL loader, Oracle and Unix Shell Scripting.
- Identifying, prioritizing of root cause of problems and resolve break and bug fixes
- Single point of contact for all the RTS issues and provide the necessary details
Confidential
Oracle Developer
Responsibilities:
- Created database triggers, procedures & Functions using PL/SQL
- Design and implementation of Stored Procedures to combine complex queries and business logic there by improve performance
- Developed UNIX shell scripts/batch jobs to insert the reader data into oracle database
- Extensively written UNIX Shell scripts to extract the meta data from flat files to a specific format and import into Oracle database using UTL FILE
- Updated the dynamic Mantas Kiosk pages by using Unix Shell script
- Worked on Control-M jobs editing and enhancement
- Extensively written UNIX Shell scripts, to extract the meta data from flat files to a specific format and import into Oracle database using SQL*loader
- Involved in User acceptance testing and Unit Testing for all the release.
- Documentation of all modules for user level and system level maintenance
- Handling production support activities on severity basis
- Analysis of the existing code and preparing design documents for enhancements. Gathering requirements from the client and doing bug fixing / enhancement of the existing code as per Client’s requirement
- Handled the Mantas application tickets
- Single point of contact for token administration and dual authentication applications which is used to secure the mantas application.
- Periodically solved the actuate server issue with help of mantas team.
- Managing the token administration and Dual authentication application which developed in JSP and Java
- Automated the daily task by using Unix Shell scripting
- Applied the mantas product patches which are one of the critical tasks in AML.
- Provided the solution for Mantas alert problem