Senior Etl/elt Developer Resume
Phoenix, AZ
PROFESSIONAL SUMMARY:
- Over 7 years of IT experience in Design, Development, Implementation, Data Migration, and Testing of Database/Data Warehousing applications using Extraction, Transformation, Loading and Data Analysis Techniques.
- Exclusive experience in Big data Technologies and Hadoop ecosystem's such as HDFS, MapReduce, PIG, Hive, Hbase, Sqoop and Flumes.
- Drawing on experience in all aspects of analytic/data warehousing solutions (Database issues, Data modeling, Data mapping, ETL Development, metadata management, data migration and reporting solutions) I have been key in delivering innovative database /data warehousing solutions to Telecom Industry & Health Care.
- Extensive experience in Pig Scripting and analyzing data using HiveQL, Pig Latin, and HBase.
- Experience in importing and exporting data using Sqoop from HDFS/Hive/HBase to Relational Database Systems.
- Worked in job workflow schedulers and monitoring applications such as Oozie and Zookeeper.
- Extensive hands on experience in using the Data warehousing tool IBM Web Sphere Data stage 7.5/ 7.5.1 Info sphere Datastage 11.3, 8.7, Talend Integration ETL, Informatics Power Center 9. worked on a modern data pipeline, to collect, cleanse and process diverse data from disparate sources across the organization.
- Having Experience in implementing data warehouse solutions, ETL, Analytic and Reporting.
- Extensively involved in the development of ETL process for extracting data from different data sources, data transformation and loading the data into data warehouse for analytical purpose.
- Strong understanding of Data Modeling (Relational, dimensional, Star and Snowflake Schema), Data analysis, implementations of Data warehousing using Windows and UNIX.
- Involved in performance tuning of targets, sources, mappings, and sessions.
- Migrated Data Stage Business intelligence tool from Version 7.x to 8.x.
- Translated business requirements into technical design specifications using Visio to design ERD schema.
- Worked with and extracted data from various databases sources Teradata, Oracle 11g/10g/9i, Sequential files and worked with different format of file systems.
- Widely worked with Teradata utilities like BTEQ, Fast Export, Fast Load, Multi Load and SQL LOADER scripts to export and load data to/from different source systems including flat files.
- Hands on experience using query tools like TOAD, SQL Developer, PLSQL developer, Teradata SQL Assistant and Query man.
- Excellent working experience on multiple platforms like LINUX, UNIX.
- Migrated DataStage 8.7 ETL to Talend Integration ETL with Hadoop
- Migrated Legacy File processing system called Daytona an Confidential & Confidential Product to Data stage 8.7 parallel editions.
- Experience in Performance tuning of ETL processes and datastage jobs in SMP and MPP environments from both system as well as job design perspective.
- Worked on complex UNIX shell scripting like cron scripts, sftp, C:D Scripts, Complex file validation scripts, Purge logic scripts and closely worked Confidential & Confidential systems leads.
- Played significant role in various phases of project life cycle, such as requirements definition, functional & technical design, testing, Production Support and implementation
- Excellent Good organizational skills, outgoing personality, Self - motivated, hardworking, ability to work independently or cooperatively in a team, eager to learn, ability to grasp quickly.
- Excellent team member with problem-solving and trouble-shooting capabilities, Quick learner, highly motivated, result oriented and an enthusiastic team player.
- Worked On the XML transformation capabilities in IBM InfoSphere DataStage 8.5 Transform XML using the DataStage XML INPUT & OUTPUT stage.
- Achieved awards like Star performer for migrating the Daytona technology to ETL Data stage 8.7v. Also honored with Pat-On-The-Back for playing significant role in various phases of project.
TECHNICAL SKILLS:
ETL Tools: IBM DataStage 11.3, 8.7, 8.5, Talend Integration ETL Tool 5 & 6, Ascential DataStage 7.5, Informatica Power Center 9
Databases: Oracle 9i/10g/11g, MS Access, CASSANDRA, MONGO DB
Languages: UNIX SHELL, HSQL, CQL (CASSANDRA QUERY), TERADATA, SQL, Confidential & Confidential Product- Cymbel Query Language, JAVA, XML
Hadoop: Hbase, Pig, Hive,Ambari
Operating Systems: Windows XP/NT/2000, UNIX, Linux.
Scripting: Unix Shell Scripting. TWS Composer Files
Other Tools: MS Office, SQL*Plus, TOAD, SQL Developer, Teradata, Management Studio.
Schedulers: Datastage Internal Scheduler, Cron Tab, Zera, IBM Tivoli TWS
PROFESSIONAL EXPERIENCE:
Confidential, Phoenix, AZ
Senior ETL/ELT Developer
Responsibilities:
- Analyze business requirements and created technical specification document for source to target mapping for the ETL development.
- Develop UML Diagrams: Use Cases, Activity Diagram, Sequence Diagram, and Class Diagram.
- Importing and exporting data from FACETS UNIX server into HDFS.
- Involved in AGILE requirement gathering and legacy system analysis.
- Based on User stories in RALLY, Estimated Development Tasks in RALLY ( CA CENTRAL AGILE) for each Sprint User Story .
- Writing UNIX file Validation, File Watcher Scripts and Pig scripts to process the data and load date in HDFS.
- Experienced in analyzing data with Hive and Pig.
- Experienced in managing and reviewing Hadoop log files.
- Experienced in defining job flows.
- Extracted the data from oracle into HDFS using Sqoop.
Environment: s: Hortonworks Hadoop, MapReduce, HDFS, Hive, Sqoop, Pig, Oozie, Linux, Shell Scripting, Oracle 12 c database, SFTP, File watcher UNIX Scripts. CA RALLY Agile central, Cron tab
Senior ETL/ELT Developer
Confidential
Responsibilities:
- Analyze business requirements and created technical specification document for source to target mapping for the ETL development.
- Importing and exporting data into HDFS and Hive using Sqoop.
- Experienced in analyzing data with Hive and Pig.
- Writing Pig scripts to process the data.
- Developed a data pipeline using Talend Integration ETL to store data into HDFS and performed the real-time analytics on the incoming data.
- Developed PIG Latin scripts to extract the data to load into HDFS.
- Involved in HBase setup and storing data into HBase, which will be used for further analysis.
- Experienced in managing and reviewing Hadoop log files.
- Experienced in defining job flows.
- Involved in creating Hive tables, loading with data and writing Hive queries using the HiveQL which will run internally in the map-reduce way.
- Extracted the data from oracle into HDFS using Sqoop.
Environment: s: HortonworksHadoop, MapReduce, HDFS, Hive, Sqoop, HBase, Pig, Oozie, Talend Integration ETL Tool, Linux, Shell Scripting. CA RALLY Agile central
Senior ETL Developer
Confidential
Responsibilities:
- Analyze business requirements and created technical specification document for source to target mapping for the ETL development.
- Involved in preparing high level and detailed design documents and acceptable differences documents for the end users.
- Data extraction from fixed width files and transformed as per the business requirement and loaded into the staging Oracle FACETS 5.2 Tables.
- Involved in AGILE requirement gathering and legacy system analysis.
- Based on User stories in RALLY, Estimated Development Tasks in RALLY ( CA CENTRAL AGILE) for each Sprint User Story .
- Recurrent billing payment for Confidential Medicade claim auto monthly billing payments.
- Worked extensively with Dimensional modeling, Data migration, Data cleansing, ETL Processes for data warehouses.
- Developed Slowly Changing Dimension Mappings for Type 3 SCD .
- Formulated UNIX File watcher and File validation Script.
- Developed Tableau workbooks from Facets Data Source for Exception and Data Balancing & Control report.
- Integrated a wide variety of source file layouts into the data warehouse.
Environment: RALLY, IBM datastage 11.5, FACETS 5.2, Oracle Developer, Oracle 12c database, UNIX Shell Scripting, IBM TIVOLI, Tableau
Senior ETL Developer
Confidential
Responsibilities:
- Collaborated with multiple health plans to understand and implement detailed requirements for HEDIS and other reporting needs.
- Created detailed functional and technical design documents.
- Planned, coordinated analysis, design and extraction of encounter data from multiple source systems into the data warehouse relational database (Oracle) while ensuring data integrity.
- Developed, documented and validated complex business rules vital for data transformation.
- Enhanced and expanded the encounter data warehouse model through sound detailed analysis of business requirements.
- As a Dev Lead appeared in AGILE SCRUM meetings and recommend plan of execution to meet business requirements.
- Involved in AGILE requirement gathering and legacy system analysis.
- Based on User stories in RALLY, Estimated Development Tasks in RALLY ( CA CENTRAL AGILE) for each Sprint User Story .
- Worked on, CQA Sprint CTO (Code Turn Over) deliveries.
- Datastage environment and Database Table initial setup for NEW AGILE processes.
- AGILE development Supported for CQA Defects and Clarification
- Designed and developed mappings between sources and operational staging targets in provided data models and data maps (extract, transform and load analysis) of the data marts for systems in the aggregation effort.
- Worked extensively with Dimensional modeling, Data migration, Data cleansing, ETL Processes for data warehouses.
- Integrated a wide variety of source file layouts into the data warehouse.
- Extensive experience in design and development of Decision Support Systems (DSS).
- Developed parallel jobs using different processing stages like Transformer, Aggregator, Lookup, Join, Sort, Copy, Merge, Funnel, CDC, Change Apply and Filter.
- Used Enterprise Edition/Parallel stages like Datasets, Change Data Capture, Row Generator and many other stages in accomplishing the ETL Coding
- Familiar in using highly scalable parallel processing infrastructure using parallel jobs and multiple node configuration files.
- Experienced in scheduling Sequence and parallel jobs using DataStage Director, UNIX scripts and scheduling tools.
- Experience in troubleshooting of jobs and addressing production issues like data issues, ENV issues, performance tuning and enhancements.
Environment: RALLY, IBM datastage 11.5, FACETS 5.2, Oracle Developer, Oracle 12c database, UNIX Shell Scripting, IBM TIVOLI, FACETS WRAPER UNIX SCRIPTS, MAINFRAME JCL
Tech Lead/Senior ETL Developer
Confidential
Responsibilities:
- Responsible for building scalable distributed data solutions using Hadoop.
- Worked hands on Datastage 8.7 ETL migration to Talend Studio ETL process.
- Design, develop, validate and deploy the Talend ETL processes for the DWH team using HADOOP (PIG, HIVE) on Hadoop.
- Collaborate with the Data Warehouse team to design and develop required ETL processes, performance tune ETL programs/scripts.
- Handled importing of data from various data sources, performed transformation like ETL (Extract Transform and Load) and ELT (Extract Load and Transform) into HDFS.
- Extracted the data from Oracle 12c, transformed and load in HDFS using Talend Studio ETL TOOL.
- Analyzed the data extraction by performing Hive queries and running Oracle SQL to know user behavior.
- Continuous monitoring with ADMIN and managing the Hadoop cluster through Cloudera Manager.
- Developed Hive queries to process the data and generate the data cubes for visualizing
Environment: Hadoop 2.2, MapReduce, HDFS, Hbase, Hive, TALEND ETL TOOL for DWH.
Tech Lead/Senior ETL Developer
Confidential
Responsibilities:
- Performing impact analysis for custom database objects, data and its dependencies in Facets core and custom tables, custom batch and front end extension etc.
- Infrastructure management performance tuning, Facets install and configuration, batch configuration etc. Data and Object migration - Custom object migration, Data migration of facets and custom tables from Sybase to Oracle 12c.
- Remediation - Technical design, development, unit testing of impacted components, upgrade current Data Stage Version 7.x and 8.5 to version 8.7/8.5 to be compatible with Oracle 12c RAC, Technical configuration, Batch configuration, Defect tracking and fixing during SIT and UAT.
- Closely worked on all datastage processes from Provider, Membership & Finance by formulating extraction, and transformation and load schemes for all process.
- Involved in business requirements gathering meetings and created functional, technical specification documents and source to target mapping documents.
- Development of data stage Job remediation from Sybase to Oracle stage design, execution, testing and deployment on the client server.
- Worked on Common Frame Work Setup, Datastage Environment Setup,
- Worked on IBM Tivoli scheduler for Data stage Schedule. Created TWS Job streams and Vartables for data stage jobs.
- Driven on Data stage environment setup like data stage node configuration, NLS setup for Data stage Oracle 12 C readable.
- Extensively worked with DataStage Designer for developing various jobs in formatting the data from different sources, cleansing the data, summarizing, aggregating, transforming, implementing the partitioning and sorting methods and finally loading the data into the data warehouse.
- Worked on complex UNIX shell scripting like MS SEND & MS GET FTP scripts, Complex file validation scripts, Purge logic scripts, Sqlloader Control files.
- Also worked in SQLLOAD Utility Scripts.
- Extensively did the Data Quality Checks on the source data.
- Operated on Confidential possessed environment like Common ETL Framework, Data stage job run through ITG, FAST Process for Deployment.
- Segmentation of data according to the categories provided by the client.
- Developing components and functions for data masking, encoding and decoding.
- Performance tuning and resolving scratch disk memory errors.
- Optimizing the search function and creating custom search patterns.
- Achieved awards achieved like Spot performer.
Environment: IBM Data stage Enterprise Edition 8.7(Data stage, Quality Stage), Oracle 11g, Fixed width files, FACETS 5.2, 4.71 Windows XP, UNIX (Shell Scripting).
Confidential, Chicago, IL
Senior Datastage ETL Developer
Responsibilities:
- Analyze business requirements and created document for the source to target mapping for the ETL development. Involved in preparing high level and detailed design documents and acceptable differences documents for the end users.
- Worked On the XML transformation capabilities in IBM Infosphere DataStage 8.5 Transform XML using the DataStage XML INPUT & OUTPUT stage
- Schedule Datastage jobs using Zena scheduling tool.
- Extracted Data from fixed width files and transformed as per the requirements and loaded into the staging Teradata Database.
- Created Datastage Parallel jobs using Designer and extracted data from various sources, transformed data according to the requirement and loaded into target databases like Teradata Database.
- Extensively worked with DataStage Designer for developing various jobs in formatting the data from different sources, cleansing the data, summarizing, aggregating, transforming, implementing the partitioning and sorting methods and finally loading the data into the data warehouse.
- Extensively did the Data Quality Checks on the source data.
- Worked with Oracle Connector and Enterprise, Peek, Dataset, Lookup, File Set, Filter, Copy, Join, Remove Duplicates, Modify, Surrogate Key Generator, Change Capture, Funnel stages.
- Involved in Integration testing, Co-ordination of the development activities, production support and maintenance of ETL Jobs.
Environment: IBM Data stage Enterprise Edition 8.7(Data stage, Quality Stage), Oracle 11g, Fixed width files, Windows XP, UNIX (Shell Scripting).
Confidential, Chicago, IL
Senior ETL Developer
Responsibilities:
- Involved in Complete Software Development Lifecycle Experience (SDLC) from Business Analysis to Development, Testing, Deployment and Documentation.
- Used Teradata utilities fast load, multiload to load data.
- Wrote BTEQ scripts to transform data.
- Wrote Fast export scripts to export data.
- Wrote, tested and implemented Teradata Fast load, Multiload and Bteq scripts, DML and DDL.
- Constructed sh shell driver routines (write, test and implement UNIX scripts).
- Wrote views based on user and/or reporting requirements.
- Involved in migration projects to migrate data from data warehouses on Oracle/DB2 and migrated those to Teradata.
- Performance tuned and optimized various complex SQL queries.
- Wrote many UNIX scripts.
- Good knowledge on Teradata Manager, TDWM, PMON, DBQL, SQL assistant and BTEQ.
- Gathered system design requirements, design and write system specifications.
- Excellent knowledge on ETL tools such as Datastage ETL, Agile team interaction.
- Worked on data warehouses with sizes from 30-50 Terabytes.
- Coordinated with the business analysts and developers to discuss issues in interpreting the requirements.
Environment: MS DOS, UNIX, Windows NT/ XP and Linux, COBOL, JCL, JAVA, SQL, PL/SQL, Teradata Macros, BTEQ, MLOAD, FASTLOAD, FAST EXPORT, Shell scripting, Teradata SQL Assistant, Teradata Manager, PMON, Putty.
Confidential
Senior ETL DeveloperResponsibilities:
- Involved in business requirements gathering meetings and created functional, technical specification documents and source to target mapping documents.
- Involved in the entire life cycle from design, developing, testing using Data Stage 8.7 Designer to develop Parallel Jobs for Extracting, Cleansing, and Transforming, Developed Shell Scripts for file validation and for data loading procedures.
- Performed DQ checks as per the requirement, which include account number validations.
- Sequential file stage, transformer and Oracle connector stages to load the data as per the requirement.
- Created Job sequencers to run the data stage jobs.
- Created extracts as per the end users request from Oracle tables and delivered to end users through SFTP.
- Involved in Development activities, coordinating with testing team and production issues.
Environment: IBM Data stage Enterprise Edition 8.7(Data stage, Quality Stage), Oracle 11g, Fixed width files, Windows XP, UNIX (Shell Scripting).
Confidential
Senior ETL DeveloperResponsibilities:
- Analyze business requirements and created document for the source to target mapping for the ETL development. Involved in preparing high level and detailed design documents and acceptable differences documents for the end users.
- Extracted Data from fixed width files and transformed as per the requirements and loaded into the staging Oracle tables.
- Created Datastage Parallel jobs using Designer and extracted data from various sources, transformed data according to the requirement and loaded into target databases like Oracle 10g.
- Extensively worked with DataStage Designer for developing various jobs in formatting the data from different sources, cleansing the data, summarizing, aggregating, transforming, implementing the partitioning and sorting methods and finally loading the data into the data warehouse.
- Extensively did the Data Quality Checks on the source data.
- Used Datastage Designer for creating new job categories, metadata definitions, and data elements, import/export of projects, jobs and datastage components, viewing and editing the contents of the repository.
- Worked with Oracle Connector and Enterprise, Peek, Dataset, Lookup, File Set, Filter, Copy, Join, Remove Duplicates, Modify, Surrogate Key Generator, Change Capture, Funnel stages.
- Involved in Integration testing, Co-ordination of the development activities, production support and maintenance of ETL Jobs.
- Involved in scheduling the Datastage jobs using Crontab.
Environment: InfoSphere Information server 8.7(Datastage, Quality Stage, Information analyzer, Designer, Director, Fast Track), Oracle 10g, Flat Files, Shell Scripting, SQL.