We provide IT Staff Augmentation Services!

Etl Senior Developer Resume

4.00/5 (Submit Your Rating)

Atlanta, GA

PROFESSIONAL SUMMARY:

  • 10.3 years of extensive experience in Data Warehouse applications using Informatica Powercenter 8.x,9.5, 9.6, 10 Grid, Oracle 10G, Teradata 14.1, 15.1, DB2, MS SQL server, IBM DB2 and UNIX platforms and BI Reporting tools Microstrategy 9.2.1, 9.0.1, 8.6 and Data Modeling, Design, and in several domains including Air lines, Healthcare & Pharmaceutical, Automobile Trading, Banking & Credit Cards and Retail.
  • Certified Microstrategy professional with over 10+ years of IT experience involved in various stages of the System Development Lifecycle (SDLC) from requirements gathering analysis, modeling, case diagrams, documentation, implementation including maintenance and support.
  • Strong Proficiency in Databases like Teradata, Oracle, PL/SQL, DB2, SQL Server and Netezza.
  • Experience in UNIX shell scripting and Teradata SQL.
  • Preparing ETL mapping Documents for every mapping and Data Migration document for smooth transfer of project from development to testing environment and then to production environment.
  • Run data quality specific ETL jobs (address standardization and validation, email cleanups, name cleanup, parsing, etc.) utilizing IDQ and other ETL tools
  • Experienced in Informatica cloud for data source from Salesforce.com and create source and target connections in Informatica cloud and synchronize data in SFDC.
  • Implemented Innovative ideas like Java transformation to convert Local timestamp to GMT timestamp and to create GUID using Java snippet, UNIX scripting to find number of semicolons (delimiter) exceeding and sending an email notification to the source team.
  • Worked on Data Profiling using IDQ - Informatica Data Quality to examine different patterns of source data. Proficient in developing Informatica IDQ transformations like Parser, Classifier, Standardizer and Decision.
  • Worked on Informatica B2B component such as Parser, Mapper, Serializer, XMAP concepts and intake of several input formats like HIPAA 837 EDI format and NCPDP libraries in Healthcare Industry layouts.
  • Provided design recommendations and thought leadership to sponsors/stakeholders that improved review processes and resolved technical problems.
  • Undergone in Hadoop, Big data fundamentals, HIVE, Hbase, Spark and Scala programming. Participated in Hadoop POC implementation
  • Worked on POC in Hadoop, HIVE and Scala programming and Informatica Data Lake using Hive tables.
  • I possess versatility to adapt to new technologies, analytical and reasoning skills, Also a quick learner and a good team player.
  • Exceptional ability to quickly master new concepts and capable of working in group as well as independently with excellent communication skills.

TECHNICAL SKILLS:

Hardware / Platforms: Windows 98/XP/2000/7, NT, UNIX, LINUX

ETL Tools: Informatica Power Center 7.1.1, 8.x, 9.5, 9.6.1, 10.0, 10.2, Datastage 6.0, Informatica Cloud, Informatica Data Quality (IDQ) 9.6.1,10.0, 10.2, Informatica Product 360, Informatica B2B 10.2, Salesforce.com

OLAP/Reporting Tools: Microstrategy- 8.1.2, 9.0.1, 9.2.1, 9.3.1, Microstrategy Certified Developer 2011, Tableau V 10, OBIEE, Hyperion Interactive Reporting Studio.

Big Data Ecosystem: Hadoop, Mapreduce, HDFS, HBase, HIVE, Scala programming. Sqoop.

Programming Languages: C/C++, Java, PL/SQL.

Scripting Languages: Unix, Shell scripting, Linux, Python

Databases: Teradata v2R6, V13.11, V14.1, V15.1, Oracle 10G, 11i, IBM Red Brick, DB2, Netezza 7.2.x, SQL Server, HBase, PostgreSQL

RELEVANT PROJECT EXPERIENCE:

Confidential, Atlanta, GA

ETL Senior Developer

Responsibilities:

  • Perform data analysis including data validation, data standardization.
  • Data profiling to review the source data, understand structure, content and interrelationships.
  • Assess the quality of the data for model development as well as inputs to the model, providing recommendations to improve the data quality at the source.
  • Uncover data quality issues using IDQ in data sources and define what needs to be corrected to standardize or remove inconsistencies in input data.
  • Work with Data governance team to define the validation rules.
  • Identify unanticipated business rules, hierarchical structures.
  • Worked on Informatica Power Center tools- Designer, Repository Manager, Workflow Manager, and Workflow Monitor.
  • Used Debugger to test the data flow and fix the mappings.
  • Works on Conceptual, Logical, Physical and Dimensional Data Modeling.
  • Perform Data Completeness analysis, Uniqueness analysis, Value Distribution analysis, Range analysis and Pattern analysis.
  • Extensively used Source Qualifier Transformation to filter data at Source level rather than at Transformation level. Created different transformations such as Source Qualifier, Joiner, Expression, Aggregator, Rank, Lookups, Filters, Stored Procedures, Update Strategy and Sequence Generator.
  • Worked on Audit process and Error process which involves IDQ hub and MDB hub.
  • Analyze project data to determine specifications or requirements.
  • Worked on Parser, serializer, XMAP for EDI file formats using Informatica Data Processor transformation.
  • Worked with Informatica GCS support on the performance issues and other B2B related library issues and IDQ issues.
  • Used B2B libraries for HIPAA 837 file format and NCPDP file format and serialize and load it into Oracle tables
  • Worked on end to end solution for the Data Processor transformation and import to PowerCenter as mapping and run through PowerCenter.
  • Utilize “Informatica Developer” 10.2 and “Informatica Analyst” to design and develop custom objects and rules, Profiling, data tables and create/import/export mappings to PowerCenter
  • Used Address Doctor and standardize rules based on the business rules using IDQ
  • Loaded data to Staging layer and cleanse data for IDQ hub and provide staging layer for MDM to master the data.
  • Build deployment plan, rollback model and migration document template and carry out deployment activities with the help of Change management team.
  • Use JAMA tool to track the Requirements and other details about the project using the agile methodology.
  • Project methodology is on the Agile Methodology by explaining on Sprint Planning, Sprint execution, Retrospective, etc.
  • Worked on SQL developer and Toad on Oracle 10g.

Confidential, Chicago, IL

ETL Senior Developer

Responsibilities:

  • Build one set of data result by correlating key paper tests to processing historian tags. The correlation periods should be metadata driven, with time or distance-speed-based offset taken directly from user input.
  • Normalize the historian tags for each plant data and correlate with MDIS data.
  • Create command tasks to run the shell script for calling PL/SQL procedures.
  • Validate the data for each column using the procedure and functions before loading into the final Oracle analytics data mart.
  • Use the cursor inside the PL/SQL procedure for validating the rules applied on each column based on the user’s p for all the Manufacturing plants data.
  • Use concurrent workflow run at the worklet and workflow level to run the historian data and MDIS data for 18 Manufacturing plants at the same time.
  • Create the informatica job to load the historical run and incremental load using the parameters passed through the parameter files.
  • Work on the score predictive model to run the job to apply unit-of-measure calculation on each plant.
  • Use informatica cloud tool to connect to Salesforce.com to bring the historian tags files from SFDC.
  • Worked on informatica cloud source and target connections and synchronize data in SFDC.
  • Use JIRA tool to track the user stories and track the project blockers and other details about the project using the agile methodology.
  • Profile source data and determine all source data and metadata characteristics; Design and execute a Data Quality Audit/Assessment; Design and execute the data quality mappings that will cleanse, de-duplicate, and otherwise prepare the project data
  • Utilize “Informatica Developer” and “Informatica Analyst” to design and develop custom objects and rules, data tables and create/import/export mappings
  • Developed and tuned Hadoop Map Reduce programs to analyze the data, populate stage tables and store the refined data into partitioned tables in the EDW.
  • Creating Hive external tables and partitioned tables using Hive Index and used HQL to make ease of data analytics.
  • Move the code the QA environment for UAT and deploy the code in Production
  • Schedule the informatica job using Appworx scheduler to run every day and Apex job to run every 15 mins.
  • Apex screen to be created for user interface to add the validation rules for each column for analytics.
  • Change request for moving the code to Production using HP service manager.
  • To create support documents for and provide support during failure in production.
  • Build one set of data result by correlating key paper tests to processing tags. The correlation periods should be metadata driven, with time or distance-speed-based offset taken directly from user input.
  • Connect to Manufacturing plant Ignition node and extract data values from ABB data files for a specified time and map or rename the tags to specified standard columns, using mapping rules defined by user via APEX.
  • Apply unit-of-measure conversions through PL/SQL procedures and packages and functions through informatica command task.
  • Use audit control tables to track the logs and errors in Procedures and workflow control tables to log the workflow and session logs.
  • Purge the previous workflow logs in the workflow and event control tables after the successful run of the workflow.
  • Use concurrent workflow or worklet run for each plant and stores in different tables for which runs at the same time for performance improvement.
  • Create SAS program to load some of the calculated fields for analytic purpose.
  • Perform bounds-checking validation rules that are stored as upper and lower bounds by target column, and takes specified actions upon validation failure that may replace the invalid column value or reject the entire row.
  • Build a complete record representing the process settings and measurements applicable to a piece of paper coming off the line.
  • Source in up to date raw material costs for Otsego from enterprise data warehouse.
  • ETL programs are orchestrated into a workflow that applies the above rules, connection information, and logical set of source types (Historian tags + relevant MDIS source tables) in the correct order and is callable from an external scheduler via a Linux shell script.
  • With minimum parameter change, need to rerun ETL workflow to reload historical data. Same workflow to be configured and scheduled to run at regular basis for incremental update.
  • Use HP Service manager to create change request for moving the code to SI and Production environment.
  • Use Project version Control to track all the versions of the code before deploying the code in Production environment.
  • Test and deploy all code objects to production environment.
  • Support and fix and bugs and issues.
  • Identify product related data in the Datawarehouse for the Confidential products and consolidate all the products information and send it to Product Information Management maintained by USG.
  • Build informatica ETL jobs to create XML files for different Confidential products using the calculation from (PIR) Product information database rules.
  • Build a batch job to send a nightly file to the Product Information Management (Product 360) tool that maintains the client product website.
  • Use Version control tool for Informatica migration between Dev, QA and Prod such as Serena PVCS Version Manager tool.
  • Document requirements and capture business rules in JIRA stories and work in an AGILE project methodology.
  • Build new Oracle PLSQL packages for automation of running the Information workflow through command task and capturing the logs in the log table.
  • Created Primary unique indexes for Primary key and unique indexes for performance improvement in the informatica jobs.
  • Worked on the real-time data and the Informatica ETL jobs capture CDC and real-time data.
  • Prepare documentation of the project design and model based on the standards followed by the client.
  • Automated process for capturing audit process and exception handling through PLSQL packages which has stored procedures, cursors and functions.
  • Used informatica Data Quality tool IDQ for performing dynamic mapping and setting rules for standardization.
  • Based on the calculations done at the ETL tool will be used by the sales reps to maintain the new Confidential values.
  • Worked on the user control reports like Apex development and OBIEE reports.
  • Collaborate with the quality assurance team in developing testing plans and streamlining testing activities
  • Used HP QC to upload Test plan and run test cases and capture defects.
  • Test and deploy all code objects to production environment.
  • Support and fix and bugs and issues.

Confidential

ETL Developer, Onsite Lead

Responsibilities:

  • Capture requirements from Functional design and prepare technical design documents and build the data model to integrate the data into the Datawarehouse systems.
  • Design and Develop data sourcing ETL jobs utilizing Informatica, Linux shell scripting, FTP connections and send XML files using web services to Product Information Management systems.
  • Used PIM tool i.e., Informatica Product 360 to create new products for different Wallboard's and use MDM tool to track the individual product business glossary into the system.
  • Complex logic needs to be built in the ETL jobs using Informatica based on the data from the PIR database and integrate with the product information from web methods and send XML files which has the calculated measure of Average pre-consumer and post-consumer recycled content of the product and send back to the Product Information Management system for the redesigned website.
  • Apply unit-of-measure conversions when needed
  • Identify performance bottlenecks and data dependencies within the data warehouse and resolve any data anomalies.
  • Work in an AGILE project methodology and utilize JIRA tool using Kanban approach to create and track user stories.
  • To give daily status reports on a scrum meeting and stand up call.
  • Support and resolve any data/job issues in production environment.
  • Tune the performance of the sessions in Informatica Grid by using pushdown optimization, persistent cache, and bulk load, table level partitioning and indexing, and writing PL/SQL packages procedures and cursors.

Confidential, Atlanta, GA

ETL Developer, Onsite Lead

Responsibilities:

  • As the SME, involved in design and Coding of scripts for extracting and loading data into data marts using one or more of the following design approaches:
  • Develop BTEQ SQL and MLoad scripts optimized for Teradata environment and executing via a scheduling tool i.e., ESP job scheduling (mainframe).
  • Design of Informatica ETL processes for loading data into data marts.
  • Worked on TPT utility and Fast load scripts for full refresh load on a daily job, monthly jobs
  • Design and build out of UNIX shell scripts as needed.
  • Worked on the Performance tuning in Informatica by using pushdown optimization, persistent cache, and bulk load.
  • Created proper PI taking into consideration of both planned access of data and even distribution of data across all the available AMPS.
  • Used datamover jobs to copy the tables from Production to History database for archival of database for history.
  • Created different parameter files and changed Session parameters, mapping parameters, and variables at run time.
  • Design and build out of ESP scheduling/Control-M for scheduling one time and recurring executions of the Informatica workflows or SQL load scripts.
  • Used SQL Execution Plan to understand Teradata query performance, tuned query to achieve greater performance.
  • Worked on gathering reporting and analysis requirements, documenting the report specifications, implementing the metadata layers including Physical, Business Model and Mapping, and Presentation layer.
  • Developed interactive Dashboards and Reports in Oracle Interactive Reporting suite as per the reporting specifications.
  • Also took part as a BA role in one of the project to gather requirements and convert it into technical documentation and provide scorecards to give Project plan, resource plan and timelines.
  • Worked on the Production Support team to monitor almost 3000 EDW jobs in the Informatica monitor as well as the ESP scheduler i.e., Mainframe application to see the dependent jobs complete on time and trigger the child jobs.
  • Collaborate with various business and technical teams to gather requirements around data quality rules and propose the optimization of these rules if applicable, then design and develop these rules with IDQ
  • Modified SQL queries to use volatile tables, inner join queries to achieve great performance.
  • Used Flat file, Oracle, SQL server, Netezza, Salesforce.com and SAP database as the source for the projects.
  • Used Informatica External loaders for fast load, Multiload and Fastexport.
  • Used Teradata TPT relational connections from Informatica to load the data from source systems to EDW.
  • Worked on the Teradata temporal feature in maintaining the history of a table with start effective and end effective timestamp showing the Valid/invalid history records.
  • Worked on the Informatica Pushdown optimization techniques and make use of Full PDO and increase the performance time of a load job to 90 % faster than the normal execution time.
  • Created more than 100 ETL mappings in Informatica, SQL’s/ DB objects in Teradata and UNIX shell scripts to implement the business logic.
  • Enhanced existing Scripts to adhere to new EDW Standards provided by the Client in projects executed.
  • Worked on the Teradata Unity suite on two projects.
  • Worked on the agile methodologies based on Sprint releases.

Environment: Informatica Powercenter 9.6,10 Grid, Informatica Data Quality (IDQ) 9.6.1, Hyperion Interactive Reporting Studio, OBIEE, Oracle 10g, Unix Shell scripting, Salesforce.com, HP Quality center 11.0, Platinum, Netezza 7.2.x, Teradata, SQL server, Teradata Unity suite

Confidential, Atlanta

BI Developer

Responsibilities:

  • Participating in the all the SDLC process starting from Design Phase, Build Phase and Testing Phase.
  • Provide necessary functional & design documentation
  • Involved in dimensional modeling of the data warehouse to design the business process, dimension and facts.
  • Design Phase - Prepare Approach, Technical Design Documents and Release Docs.
  • Involved in preparing migration docs from one environment to other ex. SIT to UAT.
  • Fixed Bugs and identified the root cause of the bugs reported.
  • Provided business solutions in writing business logic in queries in Oracle and SQL server.
  • Coordinate deployment of project related objects to Production.
  • Fix on errors - Deployment of code and data fixes by performing root cause analysis, and application performance improvements.
  • Designed and developed database views and stored procedures.
  • Worked on the Oracle source and stored procedures.
  • Reporting based on Microstrategy Reporting services and documents and Mobile dashboards
  • Used mobile widgets like map widgets
  • Created Public Objects like Grid Report, Report Service Documents, Graph, Dashboard, Prompts, Filters, Compound Metrics and Smart Metrics.
  • Worked on the development of 18 dashboards.
  • IPAD development, Dashboard creation with complex graphs like Dip sticks graph and Needle graph and customization of the reports based on the requirement.
  • Worked on Widgets like Time-series widgets.
  • Co-ordinate development-related meetings and provide status updates.
  • Created security filters for 5 different user groups. For e.g. Each group can see their level of data as Area level managers can see only Area level data for the Product Hematology.
  • Implementation of the Project in UAT Environment.
  • Worked in support activities as Mobile administrator after the Go Live of the product.
  • Work closely with Business and client for Business requirements, Review meetings.

Environment: Microstrategy, Informatica Powercenter 8.x, Oracle 10 G PL/SQL, Stored Procedures.

Confidential, New Jersey

Technical Lead, Onsite coordinator

Responsibilities:

  • Providing Project sizing and estimation for all the dashboard development.
  • Production Support: Fixed Bugs and identified the root cause of the bugs reported, Code changes and worked on Change requests.
  • Provide necessary functional & design documentation as to how the Change request would be implemented.
  • Worked in all the SDLC process starting from Requirement Gathering, Analysis, Design Phase, Build Phase and Testing Phase.
  • Work closely with Business and client for Business requirements, Review meetings and all offshore development Activities.
  • Used SQL Execution Plan to understand Teradata query performance, tuned query to achieve greater performance.
  • Performance tuning on SQL queries to use volatile tables, inner join queries.
  • Worked on the Creation of BTEQ for complex queries, Mload scripts.
  • Defect Resolution and Analysis /Clarification and provide prompt response to job failures.
  • Used Platinum tool for validation of errors before deployment
  • Fix on errors - Deployment of code and data fixes by performing root cause analysis, and application performance improvements.
  • Identify and resolve technical problems related to performance, functionality and coordinate with end users as necessary.
  • Preparing the test cases for the developed code
  • Involved in preparing migration docs from one environment to other environments and perform System integration testing and to User acceptance testing.
  • Coordinate deployment of project related objects to Production.
  • Implementation of the Project in UAT Environment.
  • Provided Warranty support after the Go Live of the product.

Environment: Informatica Powercenter 8.x, Platinum tool, Teradata V13.11, Oracle PL/SQL, SQL server

Confidential

BI Technical Lead

Responsibilities:

  • Analysis Phase, Design Phase, Build Phase and Testing Phase.
  • Design of Informatica ETL processes for loading data into data marts and tables in Oracle, Netezza.
  • Parameterization and assigned mapping variables and changed Session parameters and variables at run time.
  • Worked on the Performance tuning in Informatica by increasing DTM buffer size or blocks, persistent cache, and bulk load.
  • Created Deployment groups for the code move between Dev and QA environment and Prod environment.
  • Created ETL mappings in Informatica, SQL’s/ DB objects in Netezza, oracle PL/SQL and UNIX shell scripts for the automation of mail notification and zipping and archiving of csv files.
  • Worked on the automation of scripts for finding the number of Semicolon exceeded in a file.
  • Created Schema Objects like Attributes, Facts, Logical Tables and Formatting the reports based on requirement.
  • Involved in creating Test cases to reports.
  • Used HP QC tool for uploading the Test conditions and manual run of Test Scripts.
  • Implementation of the Project in Production Environment.
  • Involved in different Client process in order to create Change Management Request, IMR for Object Migration from development environment to Test and Production.
  • Raising defects in HP QC tool and assign to different teams in testing phase.

Environment: Windows XP, Informatica Powercenter 8.x, Netezza, Micro strategy 9.2.1, HP Quality center, IBM DB2 9.5

Confidential

BI Technical lead

Responsibilities:

  • Implemented the Incremental loading of Dimension and Fact tables.
  • Design of Informatica ETL processes for loading data into data marts.
  • Created different parameter files and changed Session parameters, mapping parameters, and variables at run time.
  • Worked on the Performance tuning in Informatica by increasing DTM buffer size or blocks, persistent cache, and bulk load.
  • Created Deployment groups for the code move between Dev and QA environment and Prod environment.
  • Created ETL mappings in Informatica, SQL’s/ DB objects in Netezza and UNIX shell scripts for the automation of mail notification and zipping and archiving of csv files.
  • Analysis Phase, Design Phase, Build Phase and Testing Phase.
  • Used oracle stored procedure as the source for the ETL mappings and worked on XML extraction.
  • Worked on the creation of documents during design phase for low level and high level design.
  • Worked as a Business Analyst and to prepare documentation on the Requirements
  • To work closely with the Business and Data architect team and work on the analysis and feasibility of the data for the Business Model.
  • Involved in different Client process in order to create Change Management Request for Object Migration from development environment to Test and Production and handled tickets from Production environment.

Environment: Informatica PowerCenter 8.x, IBM DB2 9.5, Microstrategy 9.2.1, HP Quality center 10.0, Windows XP

Confidential

BI Technical Lead

Responsibilities:

  • Took part in Requirement Analysis and Design Phase
  • Created Schema Objects like Attributes, Facts, Hierarchies, Logical Tables and Formatting the reports based on requirement.
  • Created Application Objects like Expression Prompts, Filters, Complex Metrics and Smart Metrics, Security Filters.
  • Created 10 reports and 5 documents
  • IBM DB2 database as the backend for the Microstrategy to report on the datawarehouse tables.
  • Created object migration document which needs to be moved from Development to QA environment and Production environment.
  • Write test cases to ad-hoc reports
  • Designed test scripts and processes, and conducted unit testing on the Microstrategy objects (Unit Testing) and SIT testing.
  • Worked on HP QC tool for testing.

We'd love your feedback!