Etl Lead/developer Resume
Miami, FL
SUMMARY:
- Certified professional with over 10 Years of extensive experience in various ETL Tools (Informatica, Abinitio, Talend, and Confidential ) with backend Database work experience in Teradata/Oracle/DB2/SQL Server.
- Implementation experience in Big Data Hadoop (Hive, HDFS, SQOOP)
- Diversified work experience with major clients such as Citibank, Confidential, Confidential, BCBS (Formerly Anthem/ Confidential ), Morgen Stanley - Smith Barney, Confidential, Confidential Health care etc.
- Informatica: 9.x/8.x (Power Center Client tools - Mapping Designer, Repository manager, Workflow Manager/Monitor and Server tools - Informatica Server, Repository Server manager
- Abinitio: 3.x ( GDE 3.x, Ops Console), Talend, Confidential
- Big Data Hadoop: Implemented the Confidential concept using Hadoop Technologies like Hive, HDFS, Pig, Sqoop
- Solid experience in Data Warehousing Concepts like Ralph Kimball Methodology, Bill Inmon Methodology, OLAP, OLTP, Star Schema, Snow Flake Schema, Bus Schema, Hybrid Schema, Fact /Dimension Table, Logical /Physical /Dimension Data Modeling using Erwin, Microsoft Visio
- Extensive experience in Database technologies like Teradata, Oracle, SQL Server and DB2 which comprises Stored Procedures, Functions, Views and Triggers, Complex SQL queries, Advanced analytic SQL function
- Proficient in the Integration of various data sources with multiple relational databases like Oracle11g /Oracle10g/9i, MS SQL Server, DB2, Teradata, VSAM files and Flat Files into the staging area, ODS, Data Warehouse and Data Mart.
- Excellent work experience in Performance Tuning ( Informatica Mappings, Abinitio Graphs, Oracle, Teradata (Complex SQL, Stored procedures)
- Extensive experience in writing UNIX shell scripts and automation of the ETL processes using UNIX shell scripting.
TECHNICAL SKILLS:
ETL /DWH: Informatica 9.x, Abinitio, Talend 5.x, Confidential
Big Data Hadoop: Hive/HQL, HDFS, SQOOP, Pig
Database: Teradata(14), Oracle (11g/10g/9i), SQL Server(2012/2008), DB2 (10/9), SAP HANA(Trained)
Reporting: Business Objects 4.x, Tableau 9.x
Data Modelling: Erwin, MS Visio
Programming Languages: Java, Shell Script, SQL, PL/SQL
Scheduling Tools: Autosys, Tivoli, CA7, Zeke
PROFESSIONAL EXPERIENCE:
Confidential, Miami, FL
ETL Lead/Developer
Responsibilities:
- Developed jobs in Talend to migrate the existing batch jobs from Confidential and Informatica.
- Used different components in Talend like tmap, tmssqlinput, tmssqloutput, tfiledelimitede, tfileoutputdelimited, tmssqloutputbulkexec, tunique, tFlowTolterate, tintervalmatch, tlogcatcher, tflowmetercatcher, tfilelist, taggregate, tsort, thdfsinput, thdfsoutput, tFilterRow, thiveload.
- Development of high level data dictionary of ETL data mappings and transformations from a series of complex Talend data integration jobs.
- Developed mappings to load Fact and Dimension tables, SCD Type 1 and SCD Type 2 dimensions and Incremental loading and unit tested the mappings
- Providing expert resolution to technical/Functional problem being a Subject Matter Expert for the product.
- Developed ETL jobs to dump the data from RDBMS to HIVE.
- Worked on fine tuning the bigdata batch jobs in Talend.
- Developed Tableau reports to run on top of HIVE/HBASE tables.
- Utilized JSON Talend components to handle the complex JSONs file.
- Deal with large volumes of data around 10 billion rows (historical, pre aggregate data)
- Developed multiple MapReduce jobs using Talend for data cleaning and preprocessing.
- Work with development and test team to ensure all bugs are removed before code is deployed in UAT Participated in designing and development of software using the Agile Methodologies.
Technology Used : Talend 5.x, Confidential, Informatica, HIVE, SQOOP, HDFS, SPARK, Business Objects, Tableau 9, UNIX, Teradata, BTEQ, Stored procedures, SQL Server, Rally
Confidential, Atlanta, GA
ETL Lead/Developer
Responsibilities:
- Developed jobs in Talend to migrate the jobs from Ab-Initio.
- Used different components in Talend like tmap, tmssqlinput, tmssqloutput, tfiledelimitede, tfileoutputdelimited, tmssqloutputbulkexec, tunique, tFlowTolterate, tintervalmatch, tlogcatcher, tflowmetercatcher, tfilelist, taggregate, tsort, thdfsinput, thdfsoutput, tFilterRow, thiveload.
- Development of high level data dictionary of ETL data mappings and transformations from a series of complex Talend data integration jobs.
- Developed mappings to load Fact and Dimension tables, SCD Type 1 and SCD Type 2 dimensions and Incremental loading and unit tested the mappings.
- For reporting needs, developed the complex BTEQ scripts to load the Teradata tables.
- Responsible for Technical Design, Development, Unit Testing, Code Review, Code migration, supporting System testing, scheduling the jobs thru HDDTM.
- Wrote complex SQL queries to validate & profiling data using Advanced Analytic SQL functions
- Stored intermediate data in HDFS, queried using Hadoop tools like HIVE etc.
Technology Used : Talend 5.x, Abinitio, UNIX, Teradata, SQL Server, BTEQ, Hadoop, HIVE, JIRA, HDFS, SQOOP, IMPALA
Confidential, Atlanta, GA
Technical Lead
Responsibilities:
- Created generic design architecture for unloading data from 500 odd tables, which will also strip off the noisy data being generated from the source.
- Co-ordinated with the data modelers to create Hive tables which will replicate the current warehouse table structure.
- Lead the team to load the history data/dead data (7 years) using SQOOP.
- Created HBase tables to load large sets of structured, semi-structured and unstructured data coming from UNIX, NoSQL and a variety of portfolios
- Lead the team in creating a generic process and build psets to unload the data from DB2 and load in to Hive warehouse (Source like Stage).
- Developed simple map - reducers scripts using JAVA
Technology Used : Talend, Hive, SQOOP, Abinitio, UNIX, DB2, HDFS, HBASE, Java
Confidential
ETL Lead/DeveloperResponsibilities:
- Developed mappings to load Fact and Dimension tables, SCD Type 1 and SCD Type 2 dimensions and Incremental loading and unit tested the mappings Providing expert resolution to technical/Functional problem being a Subject Matter Expert for the Application
- Developed, tested and reviewed complex Ab Initio graphs, sub - graphs, DML, PSET, XFR, deployed scripts, DBC files for connectivity, create Package and exports
- Analyzed business requirements to identify business rules and develop technical specifications
- Created a Validation layer and response file creation system to validate and provide an automated response to the source systems indicating the errors in the data they sent over.
- Redesigned the existing graphs and documented all the new and enhancement requests.
- Monitored performance reports from database and ETL tools server log to further consider process enhancements and tunings.
- Work Order (non-Projects) life cycles: Initiation to Closure and implemented SDLC phases for all the work efforts and document phase-associated deliverables.
- Responsible for mentoring Developers and Code Review of Mappings developed by other developers
- Worked on stored procedures in DB2 which generate reports for the Business Users.
Technology Used : Abinitio, UNIX, DB2, Stored procedures, CA7, VBA scripts
Confidential
ETL Lead/DeveloperResponsibilities:
- Understanding the specifications for Data Warehouse ETL Processes and interacting with the business analysts and end users to understand system requirements
- Working closely with Business Analysts to interpret the Business rules and make necessary modifications to process the data for accuracy.
- Performed Data Analysis and created High Level and Technical Design documents based on system requirements.
- Design ETL application and develop Data warehouse applications based on the technical / functional specifications.
- Extracted data from various heterogeneous sources like Oracle, SQL Server, DB2, MS Access, and Flat Files
- Worked on the multiple projects involving the cross-platform development and testing (Mainframe, Abinitio, Unix, DB2)
- Third Party Relationship Oversight (bridge between team, Client and third party vendors) and External Network Connectivity (for data transfer)
- Support the development team in code reviews and defect tracking.
- Prepared the Detailed Design Document for the all the modules required for development
- Analyzed the issues with the unmatched records and provided code fix to the problems using the VBA scripts
Technology Used: Abinitio, UNIX, DB2, Oracle, CA7, VBA Scripts
Confidential
Sr. Informatica Developer
Responsibilities:
- Interacted with Data Modelers and Business Analysts to understand the requirements and the impact of the ETL on the business
- Developed technical specifications of the ETL process flow
- Designed and developed complex mappings by using Lookup, Expression, Update, Sequence generator, Aggregator, Router, Stored Procedure, etc., transformations to implement complex logics while coding a mapping
- Involved in writing shell scripts on Unix for Informatica ETL tool and data validation
- Used TWS to schedule the Informatica workflows.
- Used SQL queries to perform backend testing on the database
- Conducted Informatica Knowledge Sharing sessions with other team members
- Created Test plans and Test cases in all phases such as Unit, System, Integration and User Acceptance
Technology Used: Informatica Power Center Designer, Workflow Manager and Workflow Monitor, Oracle, Teradata, UNIX, TWS
Confidential
ETL Lead/DeveloperResponsibilities:
- Development and analysis of the specifications provided by the clients.
- Performed the performance evaluation of the ETL for full load cycle.
- Checked Sessions and error logs to troubleshoot problems and also used debugger for complex.
- Was responsible for production monitoring and fixing abends.
- Performed extensive system testing and migration into production
- Preparation of Test Plans and Test Cases.
- Conducted walk through on all proposed changes, interviews with users concerning all modifications and impact analysis.
- Debugging and troubleshooting any technical issues while implementing the applications
- Communicate progress, test results to the stakeholders. Status reports preparation and coordination.
- Documented and presented the production/support documents for the components developed when handing-over the application to the production support team .
Technology Used: Informatica Power Center Designer, Workflow Manager and Workflow Monitor, Oracle, Teradata, UNIX, TWS
Confidential
ETL Lead/DeveloperResponsibilities:
- Development and analysis of the specifications provided by the clients.
- Coding and Testing of modules related to my application.
- Performed extensive system testing and migration into production
- Preparation of Test Plans and Test Cases.
- Conducted code reviews developed by my team mates before moving the code into QA.
- Provided support to develop the entire warehouse architecture and plan the ETL process .
Technology Used: Informatica Power Center Designer, Workflow Manager and Workflow Monitor, Oracle, Teradata, UNIX, TWS
Confidential
Informatica developer
Responsibilities:
- Development and analysis of the specifications provided by the clients.
- Coding and Testing of modules related to my application.
- Was responsible for production monitoring and fixing abends.
- Involved in migrating Informatica ETL application and Database objects through various environments such as Development, Testing, UAT and Production environments.
- Conducted walk through on all proposed changes, interviews with users concerning all modifications and impact analysis.
- Debugging and troubleshooting any technical issues while implementing the applications
Technology Used: Informatica 8.X UNIX, Oracle
Confidential
ETL Developer
Responsibilities:
- Development and analysis of the specifications provided by the clients.
- Coordinate with the Onshore and plans the deliverables for the day
- Worked with creating Dimensions and Fact tables for the data mart.
- Worked on complex Source Qualifier queries, Pre and Post SQL queries in the Target.
- Expression, join, Aggregator, source qualifier editors were used in the Developing the Informatica workflow.
- Developed Complex SQL queries and PL/SQL Stored procedures in Teradata V2R5
Technology Used: Informatica, UNIX and Teradata.