Etl/teradata Consultant Resume
Framingham, MA
SUMMARY:
- Around 8+ years of experience in ETL (Extract Transform Load), Data Integration and Data Warehousing using Informatica, Ab Initio, Teradata and Oracle technologies .
- Around 1 year of hands on experience in Hadoop tools like Hive, Sqoop and Hbase, Map Reduce.
- Extensive experience various business domains like Healthcare, Financial, Investment and Retail
- Strong expertise in Analysis, Design, Development, Implementation, Modeling, Testing, and support for Data warehousing applications.
- Experience working with Teradata Parallel Transporter (TPT), BTEQ, FastLoad, MultiLoad, TPT, SQL Assistant, DDL and DML commands.
- Proficient in Teradata EXPLAIN plans, Collect Stats option, Primary Indexes (PI, NUPI), Secondary Indexes (USI, NUSI), Partition Primary Index (PPI), Join Indexes (JI), Volatile, global temporary, derived tables etc.
- Extensive knowledge in Business Intelligence and Data Warehousing Concepts with emphasis on ETL and System Development Life Cycle (SDLC).
- Working Knowledge of Data warehousing concepts like Star Schema and Snowflake Schema, Data Marts, Kimball Methodology used in Relational, Dimensional and Multidimensional data modelling.
- Proficient in Database Performance, SQL Query tuning
- Extensive experience in implementing slowly changing dimensions to maintain historical data and for change data capture(CDC)
- Sound knowledge in Data Migration from DB2 and Oracle to Teradata using a automated u nix shell scripting, Oracle/TD SQL, TD Macros and Procedures etc.
- Proficient knowledge in ER and Dimensional Modeling, identifying Fact and Dimensional Tables with data modeling tools ERWIN and ER Studio.
- Work on daily Admin tasks of managing Databases, Users, Tables, Indexes, statistics, managing permissions (Roles & Profiles), addressing the user issues such as resetting passwords, unlocking user ID’s etc.
- Designed and Managed BAR strategy and Architecture (scheduling, reporting). Experience in developing complex Mappings using Variables, Mapping Parameters, and Dynamic Parameter Files for improved performance and increased flexibility.
- Experience with various source and target systems like flat files, Xmls, Cobol files, Web Services etc.
- Experience in resolving on - going maintenance issues and bug fixes; monitoring Informatica sessions as well as performance tuning of mappings and sessions.
- Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems and vice-versa.
- Involved in the Teradata Database Upgrade from version 13.0 to 13.10 to 14.0 to 14.10 and coordinating the regular Production Outages.
- Worked on Data migration projects to move data from DB2 to Teradata, SQL server to Teradata, Sybase to Teradata.
- Worked on TARA GUI & Created Archive & Restore polices on BAR serves.
- Performed system Analysis and QA testing and involved in Production Support.
- Scheduling the automated jobs for daily, weekly and monthly jobs using UNIX Shell scripts for Autosys scheduling.
- Very strong in Shell Scripting (KSh, Bourne Shell) and scheduling using Crontab.
- Experience in UNIX working environment, writing UNIX shell scripts for Informatica pre & post session operations.
- Ability to solve issues by coordinating across cross-functional areas
- Experience with the day to day maintenance of application systems in operation, including tasks related to identifying and troubleshooting application issues and issues resolution or escalation.
TECHNICAL SKILLS:
Teradata Utilities: BTEQ, FastLoad, MultiLoad, TPT, TPump, SQL Assistant, Viewpoint, Query Monitor .
ETL Tools: Informatica Power Center 9. x /8.x/7.x (Source Analyzer, Repository Manager, Transformation Developer, Mapplet Designer, Mapping Designer, Workflow Manager, Workflow Monitor, Warehouse Designer and Informatica Server), Informatica Data Quality (IDQ).
Databases: Teradata 15, 14.10/14/13.10/13, Oracle 11g/10g,8i, DB2/UDB, SQL Server
Languages: SQL, PL/SQL, XML, UNIX Shell Scripting
Operating Systems: Windows 95/98/NT/2000/XP, UNIX, Linux, NCR MP-RAS UNIX
Data Modeling: Erwin, ER Studio
Tools: /Utilities: PLSQL Developer, TOAD, Hadoop, Hive, Pig, SQOOP, SQL Developer, Erwin, Microsoft Visio, Talend, DataStage,Mainframes
Scheduler: UC4, Control M, Autosys
PROFESSIONAL EXPERIENCE:
Confidential, Framingham, MA
ETL/ Teradata Consultant
Responsibilities:
- Responsible for requirements gathering for an enhancement requested by client. Involved in analysis and implementation for an Intranet Based Information Management Information System.
- Responsible for designing ETL strategy for both Initial and Incremental loads.
- Developed the Teradata Macros, Stored Procedures to load data into Incremental/Staging tables and then move data from staging to Journal then move data from Journal into Base tables
- Interacted with business community and gathered requirements based on changing needs. Incorporated identified factors into Informatica mappings to build the Data Mart.
- Writing UNIX shell scripts to support and automate the ETL process.
- Worked on Backup/archive (BAR) using TaraGUI, Netbackup.
- Provided scalable, high speed, parallel data extraction, loading and updating using TPT.
- Performed Query Optimization with the help of explain plans, collect statistics, Primary and Secondary indexes. Used volatile table and derived queries for breaking up complex queries into simpler queries. Streamlined the Teradata scripts and shell scripts migration process on the UNIX box.
- Developed UNIX scripts to transfer the data from operational data sources to the target warehouse.
- Very good understanding of Database Skew, PPI, Join Methods and Join Strategies, Join Indexes including sparse, aggregate and hash.
- Performed Configuration Management to Migrate Informatica mappings/sessions /workflows from Development to Test to production environment.
- Extracted data from various source systems like Oracle, Sql Server and flat files as per the requirements
- Used extensively Derived Tables, Volatile Table and GTT tables in many of the ETL scripts.
- Used Informatica Designer to create complex mappings using different transformations like Filter, Router, Connected & Unconnected lookups, Stored Procedure, Joiner, Update Strategy, Expressions and Aggregator transformations to pipeline data to Data Mart.
- Used Mapplets for use in mappings thereby saving valuable design time and effort
- Used Informatica Workflow Manager to create, schedule, execute and monitor sessions, Worklets and workflows.
- Generator, Update Strategy, Joiner, Stored Procedure, and Union to develop robust mappings in the Informatica Designer.
- Worked on capacity planning, reported disk and CPU usage growth reports.
- Implemented various Teradata alerts using Alert facility in Teradata View point. Involved in setting up alters to page DBA for events such as node down, AMP down, too many blocked sessions, high data skew etc.
- Used PDCR and CPPT tools for collecting facility to setup AMP usage collection, canary query response, spool usage response etc.
- Worked on Informatica Power Center tools - Designer, Repository Manager, Workflow Manager, and Workflow Monitor and IDQ.
- Performed the data profiling and analysis making use of Informatica Data Explorer (IDE) and Informatica Data Quality (IDQ)
- Extracted the data from Teradata into HDFS using the Sqoop and Exported the patterns analyzed back to Teradata using Sqoop .
- Using various transformations like Filter, Expression, Sequence Generator, Update Strategy, Joiner, Stored Procedure, and Union to develop robust mappings in the Informatica Designer.
- Extensively developed UC4 jobs to schedule Power center workflows and Data Quality workflows.
- Involved in root cause analysis, management & User communication
- Experience in Incident Management, Change Management and Problem Management
- Provide daily support with resolution of escalated tickets and act as liaison to business and technical leads to ensure issues are resolved in timely manner
- Implemented TASM Workloads, developed TASM exceptions, filters and throttles as needed basis.
- Responsible for Customer Relations and End-user Support of the assigned site.
- Backup, Archive and Recovery (BAR): Design, Implement, maintain and automated the appropriate jobs using Tara GUI and BAR servers (Net Vault).
- Suggest fixes to complex issues by doing a thorough analysis of root cause and impact of the defect
- Provided 24/7 On-call Production Support for various applications and provided resolution for night-time production job, attend conference calls with business operations, system managers for resolution of issues.
Environment: Teradata 15,14, Informatica Power Center 9.1/9.5, Workflow Manager, Workflow Monitor, Warehouse Designer, Source Analyzer, Transformation developer, Map let Designer, Mapping Designer, Repository manager, Informatica Cloud, Informatica Data Quality (IDQ), UC4, Control-M, UNIX, SSH (secure shell), TOAD, ERWIN.
Confidential, Cary, NC
ETL / Teradata developer
Responsibilities:
- Created mappings using transformations like Source Qualifier, Aggregator, Expression, lookup, Router, Filter, Update Strategy, Joiner, Union, and Stored procedure, and XML transformations.
- Worked on Informatica Power Center tools - Source Analyzer, Warehouse Designer, Mapping &Mapplet Designer, and Transformation Developer.
- Converted existing PL/SQL Packages to ETL Mappings using Informatica Power Center.
- Used Error handling strategy for trapping errors in a mapping and sending errors to an error table.
- Implemented parallelism in loads by partitioning workflows using Pipeline, Round-Robin, Hash, Key Range and Pass-through partitions.
- Worked on Exchange Management Console (EMC) or the Exchange Management Shell (EMS) related topics.
- Extensively used Informatica Power Center to extract data from various sources, which included Flat files, Sql server, Oracle, Ms-Access and XML.
- Extensively involved in performance tuning, recommending SQL queries for better performance.
- Developed in SCRUM iterations using Agile Methodology, Iterative development and Sprint Burn down with Story Boards.
- Estimates and planning of development work using Agile Software Development.
- Used Change Data Capture (CDC) to simplify ETL in data warehouse applications
- Used debugger to debug mappings to gain troubleshooting information about data and error conditions.
- Build tables, views, and created different indexes UPI, NUPI, USI and NUSI on tables.
- Wrote Teradata Macros and used various Teradata analytic functions.
- Wrote, tested and implemented Teradata Fastload, Multiload and BTEQ scripts to transform and export data.
- Query optimization using Explain plans, Collect statistics, Primary and Secondary indexes.
- Extensively worked in the performance tuning of transformations, Sources and Targets.
- Involved in the logical and physical design of the database and creation of the Database Objects.
- Involved in migration projects to migrate data from data warehouses on Oracle/DB2 and migrated to Teradata.
- Performance tuned and optimized various complex SQL queries.
- Daily duties including transferring and converting data from one platform to another to be used for further Analysis.
- Worked with high volume datasets from various sources like Oracle, Text Files and xml targets.
- Used GIT as the version control to maintain the different versions of the code in Teradata.
- Wrote many UNIX scripts to automate the tasks of running the ETL jobs.Widely used batch scripting in Unix to support ETL and BTEQ scripts.
- Documented the issues resolved during the different phases like development and post production.Worked on Migration Strategies between Development, Test and Production Repositories.
- Supported the Quality Assurance team in testing and validating the Informatica workflows.
- Did unit test and development testing at ETL level in my mappings.
Environment: Informatica Power Center 9.1( Power Center Repository Manager, Designer, Workflow Manager, and Workflow Monitor),,Teradata 14.10, Oracle 11g, PLSQL Developer, SQL, PLSQL, UNIX Shell Scripting, Autosys, Hive .
Confidential
Teradata Developer /ETL Developer
Responsibilities:
- Development of scripts for loading the data into the base tables in EDW and to load the data from source to staging and staging area to target tables using FastLoad, MultiLoad and BTEQ utilities of Teradata.
- Writing scripts for data cleansing, data validation, data transformation for the data coming from different source systems.
- Performed application level DBA activities creating tables, indexes and monitored and tuned Teradata BETQ scripts using Teradata Visual Explain utility.
- Written complex SQLs using joins, sub queries and correlated sub queries. Expertise in SQL Queries for cross verification of data.
- Developed the Teradata Macros, Stored Procedures to load data into Incremental/Staging tables and then move data from staging into Base tables.
- Performed Space Management for Perm & Spool Space.
- Reviewed the SQL for missing joins & join constraints, data format issues, miss-matched aliases, casting errors.
- Developed procedures to populate the customer data warehouse with transaction data, cycle and monthly summary data, and historical data.
- Dealt with initials, delta and Incremental data as well Migration data to load into the Teradata.
- Analyzing data and implementing the multi-value compression for optimal usage of space.
- Query Analysis using Explain for unnecessary product joins, confidence factor, join type, order in which the tables are joined.
- Very good understanding of Database Skew, PPI, Join Methods and Join Strategies, Join Indexes including sparse, aggregate and hash.
- Used extensively Teradata Analyst Pack such as Teradata Visual Explain, Teradata Index Wizard and Teradata Statistics Wizard.
- Used extensively Derived Tables, Volatile Table and GTT tables in many of the ETL scripts.
- Tuning of Teradata SQL statements using Explain analyzing the data distribution among AMPs and index usage, collect statistics, definition of indexes, revision of correlated sub queries, usage of Hash functions, etc.…
- Flat files are loaded into databases using Fast Load and then used in the queries to do joins.
- Use SQL to query the databases and do as much crunching as possible in Teradata, using very complicated SQL Query optimization (explains plans, collect statistics, data distribution across AMPS, primary and secondary indexes, locking, etc.) to achieve better performance.
- Use PMON, Teradata manager to monitor the production system during online day.
- Excellent experience in performance tuning and query optimization of the Teradata SQLs.
- Developed mappings in Ab Initio to load the data from various sources using various Ab Initio Components such as Partition by Key, Partition by round robin, Reformat, Rollup, Join, Scan, Normalize, Gather, Merge etc.
- Created checkpoints, phases to avoid dead locks and tested the graphs with some sample data then committed the graphs and related files into Repository from sandbox environment.
- Then schedule the graphs using Autosys and loaded the data into target tables from staging area by using SQL Loader.
- Implemented Data parallelism by using Multi-File System, Partition and De-partition components and also preformed repartition to improve the overall performance.
Environment: Teradata 12, Informatica 8.6/8.1(Designer, Repository Manager, Workflow Manager, Workflow Monitor), Informatica 8x, Oracle 10G, UNIX, Citrix, Toad, Putty.
Confidential
AbInitio/Teradata Developer
Responsibilities:
- Worked in the team of two and developed a module to extract data from a source table with the help of dbc file, transformed as per business logic and loaded into the warehouse with taking backup at every point needed.
- Responsible in delivering the code before deadlines and responsible to get the code approved by the client.
- Worked in Linux environment and build scripts to automate tasks like environment cleanup and archiving logs.
- Also, developed Parameter Sets to execute the AbInitio graphs from the Linux environment.
- Responsible in solving the issues raised by Testing team to fix the code and if required to enhance for the better performance.
- Worked with the different format of files like xml, flat files, tables. Also, handled real time events from CRM and OMS sources.
- Later worked in Production Support Environment to handle the real-time issues in production.
- Transformed and processed the daily delta feeds of customer’s data to various targets which is later used for reporting purpose.
- Good Knowledge on the concepts of Sand box, EME for check in&checkout process and extensively used air commands for check-in and checkouts.
- Used Teradata as the target Warehouse and performed basic operations on Teradata to verify the loaded data.
- Modified the existing procedures based on the client requirements
- Involved in performance tuning and optimized various complex SQL queries for faster user query performance.
- Created Views for various tables based on the user and reporting requirements.
- Worked under pressure to resolve the issues in very less time frame in production environment to avoid the roadblock to process production data.
- Performance Tuning, Monitoring using DB2 Tools.
- Provide DB2 Connect Support for Developers, Business users.
- Should perform RCA (Root Cause Analysis) for high level severity issues and must submit the analysis to the client.
- Also, responsible to take up the tickets raised by client in Clear Quest (QA tool) and resolved before the deadline.
- Participated in status calls every morning and must update the status to the clients.
- Created and scheduled Cron Jobs such as for Backup, System Monitoring and removal of unnecessary files.
- Updated the runbooks for various source codes for deployment and follow-up the release management.
Environment: AbInitio,Putty,Oracle,Teradata, RHEL, Cron, Confidential Clear Quest