We provide IT Staff Augmentation Services!

Sr. Etl Developer Resume

2.00/5 (Submit Your Rating)

Columbus, OH

PROFESSIONAL SUMMARY:

  • 7 years of Intelligent Data Integration and ETL experience in application development with the large Enterprise Data Warehouse and Business Intelligence systems.
  • Strong Knowledge on contemporary Data Warehousing trends - Bill Inman, Ralph Kimball methodologies, Star Schema, Snowflake schema, ODS, EDW, DM, OLAP Dimensions and Facts.
  • Analyze, validate and refine Business and Functional requirements to implement the strategic Data Warehouse Life Cycle for efficient Business Intelligence trends.
  • Develop and make enhancements to new/existing dimensional models in preparing scalable, reliable and optimized physical, logical and conceptual data models.
  • Work with Application Architects in defining project specific Blue Prints and Architectural diagrams.
  • Prepare Source to Target Data Mappings, ETL Design and technical specification documents in compliance with the data governance and best standards and review with the Enterprise Architects and ETL Admins.
  • Develop PLSQL procedures and packages that store, retrieve and manipulate data for data analytics.
  • Develop reusable code snippets, automated data movement reconciliation and Batch Framework techniques.
  • Implement Error Logging and Alert/Abort process for any data flow issues and enable restart ability/recover techniques through Informatica Power Center tool.
  • Perform Data Validation and Reconcilement checks for each data movement process and log the results for auditing purpose. Deploy the ETL/Oracle/Linux code through DevOps tools like Git and Jenkins.
  • Implement Data Retention techniques for auditing purpose and automate the Purge mechanism for very old/unused data extracts.
  • Proficient in leveraging latest databases Oracle 12c, Teradata and Netezza in achieving better performance through Partitioning, Optimizers and MPP (Massively Parallelization Processing) architecture.
  • Perform troubleshooting on production issues, identify application bottlenecks and opportunities to resolve the errors and optimize the performance.
  • Work with ETL admins in setting up Operating System Profiles (OSP), Red Hat Linux NAS/SAN locations, Domain, Node and Repository Configurations, Version Control and Deployments to higher environments.
  • Work with DBA in sizing data and index table spaces for database creation, creating and granting system privileges/roles, deploying DDL/DML scripts to higher environments.
  • Strong knowledge on Guidewire Policy Center Integration, Big Data, Hadoop File System, Financial and Insurance markets.

TECHNICAL SKILLS:

Data Warehouse: Ralph Kimball and Bill Inman Methodologies, SCD Type 1 or Type 2, Star Schema, Snow Flake Schema, Dimensions, Facts, OLTP, ODS, EDW, DM and OLAP.

Data Model: Dimensional Model and ER Models.

Databases: Teradata, Netezza and Oracle 12c.

Files: Fixed Width Flat Files, Delimited Flat Files and XML.

Languages: Teradata SQL, NZSQL, SQL, PLSQL and Shell Script

TOOLS: Data Integration: Informatica Power Center Tools (10.x or 9.x)

Data Model: Erwin and Visio

Database: Teradata Studio, AQT, TOAD, SQL Developer and SQL Plus.

FTP: Putty, Ultra Edit and WinSCP

DevOps Tools: GIT, Artifactory, Jenkins, Tortoise SVN and CA Software Change Manager

Schedulers: CA7, CICS Explorer, ESP (CA WA Workstation) and Autosys.

Testing: HP ALM Quality Center

Incident Management: Service Now, Oracle Identity Manager (OIM)

PROFESSIONAL EXPERIENCE:

Confidential, Columbus, OH

Sr. ETL Developer

Responsibilities:

  • Analyze business requirement, define development scope and develop technical specification documents.
  • Code, test, modify, debug, document, and implement ETL workflows using Informatica Power Center tools.
  • Solid experience with data modeling and database design for operational as well as analytical systems.
  • Develop scripts to automate the execution of ETL Workflows/mappings using shell scripts under UNIX environment.
  • Integrating and Ingesting all policies information to Guidewire Policy Center using ETL Informatica tools.
  • Participate in code & design reviews and ensure that all solutions are aligned to pre-defined architectural specifications; identify/troubleshoot application code-related issues; and review and provide feedback to the final user documentation.
  • Develop Batch Framework through Perl scripts which set Informatica Environment, Configure Database Connections, Trigger Informatica workflows, Update Job Status tables and create daily log files.
  • Extensive experience and LONG datatypes in writing PLSQL blocks including Stored Procedures, Functions, Triggers and Packages. Importing/exporting the policy life cycle data from/to Guidewire Policy Center.
  • Developed Change Data Capture flows for incremental loads, PLSQL blocks that perform gather stats, data archiving and purging during DB maintenance window.
  • Version and revision control practices and procedures for all ETL components including Informatica Workflows, DB scripts, Parameter files and configuration files thru Tortoise SVN and CA Software Change Manager.
  • Develop ESP Jobs, Events and applications through CA WA Workstation and CICS Explorer tools.
  • Develop Data Quality scripts that will validate the source and target systems after loading/generating the data for downstream systems.

Environment: Informatica Power Center Tools 9.6.1 Hot Fix, Oracle, PL/SQL, SQL, LINUX, Flat Files, XML, Perl, Toad, Putty, Tortoise SVN, CA Software Change Manager, CICS Explorer, ESP (CA WA Workstation), HP ALM Quality Center, Ultra Edit, WinSCP and Service now.

Confidential, Harrisburg, PA

Sr. ETL Developer

Responsibilities:

  • Understand Business, Functional Requirements and Data Mappings for ETL design and implementation.
  • Develop ETL processes using Informatica Power Center tools to implement logical and physical dataflow models according to functional requirements and technical designs.
  • Develop slowly changing Dimension (SCD Type 1, Type 2) to load Dimension Tables and implemented Change Data Capture (CDC) mechanism for Incremental Load.
  • Develop SQL/PLSQL queries and ETL (extract, transform, and load data) that is consistent and integrated with existing ETL processes.
  • Code, test, and implement Informatica mappings and scripts to build and maintain automated ETL processes.
  • Fix high priority production issues and support daily/weekly/monthly running jobs by partnering with other IT areas in resolving issues and improving processes.
  • Perform Data Quality validations for each data movement flow that comprises of critical balances and counts.
  • Provide resolution to an extensive range of complicated ETL-related problems, proactively and as issues surface.
  • Coordinate and develop process automation/job scheduling, perform testing and defect analysis.
  • Work with technical leads, architects, subject matter experts, developers, technical support groups and test teams to deliver value to the business.
  • Develops Data Quality rules and mappings based on the requirements and technical design.
  • Participate in design sessions with ETL developers, report developers, DBAs and Business Analysts.
  • Remain current on new ETL techniques and methodologies and communicate trends and opportunities to management and other developers as needed.
  • Oversees development of project plans, including requirements and level of effort (LOE).
  • Ensures communication of project status to leadership and issues within team.

Environment: Informatica Power Center 9.6/9.5 Hot Fix, Erwin, Teradata, DB2, Netezza, Oracle, PL/SQL, SQL, LINUX, Flat Files, XML, AQT, Toad, Putty, Tortoise SVN, CA SCM, CICS Explorer, ESP (CA WA Workstation), HP ALM Quality Center, Ultra Edit, WinSCP and Service now.

Confidential, CT

ETL Developer

Responsibilities:

  • Interact with business analysts and end users to understand requirements and convert business specifications to technical artifacts and determine most efficient design solution across multiple integration projects
  • Adhere to architecture standards, risk management and security policies, leading presentations in architecture review, strategic technology directions, best practice development (eg, estimating models), mentoring less experienced team members, code reviews
  • Design, analyze, develop, code, test, debug and document programming to satisfy business requirements
  • Develop Change Data Capture flows using Informatica, Oracle, Unix (Korn Shell scripts), PL/SQL blocks.
  • Thoroughly perform unit testing ensuring minimal code defects out of build phase into system test phase
  • Good communication skills and solid knowledge on software delivery lifecycle in large Warehouse environment
  • Implement Data Quality techniques for each incremental load that captures changes through Change Data Capture mechanism from the upstream systems.
  • Extensive Hands on experience designing and developing all Data Integration flows using Informatica.
  • Well versed in the following data domains: Master Data, Operational Data, Analytical Data, Unstructured Data
  • Complete code reviews for ETLs and related components, and complete documentation of issues identified and action items.
  • Correct testing defects and support all testing, including but not limited to: Development Integration Testing, System Testing, User Acceptance Testing, End-to-End Testing, and Performance Testing.
  • Provide resolution to an extensive range of complicated ETL-related problems, proactively and as issues surface.
  • Work with initiative leads, business areas, technical leads, architects, subject matter experts, developers, technical support groups and test teams to deliver value to the business.

Environment: Informatica Power Center 9.5.1, Teradata, Oracle, PL/SQL, SQL, SQL Loader, LINUX, Flat Files, Toad, Putty, Tortoise SVN, Autosys, HP ALM Quality Center and Ultra Edit.

Confidential, Hartford, CT

ETL Developer

Responsibilities:

  • Create data mapping and workflow using Informatica PowerCenter to extract, transform and load data into the target reporting environment.
  • Conform to Informatica workflows and mapping standards so that development is done in a consistent manner across all development resources (e.g., no hard coding, leverage of mapplet / reuse concept).
  • Develop SCD Type 1, Type 2 using Change Data Capture to maintain history of transactions for Incremental.
  • Tune and optimize ETL jobs for performance and throughput. Interact with data modelers and DBAs on data architecture optimization for ETL performance.
  • Provide development support, including walk through approvals, to other ETL application resources to insure that standards are followed and optimized workflow is implemented.
  • Advocate and enforce secure coding principles, as defined coding policies, to insure we are in compliance with all defined standards.
  • Assist in documenting and maintaining operational procedures in the form of an operations manual for all ETL workflows.
  • Writing Data Quality validation procedures to enable data completeness strategy for accurate information for BI
  • Assist with troubleshooting efforts for ETL job issues related to pre-production and production.
  • Leverage Data Quality as a mandatory process step for profiling data prior to downstream transformation.
  • Contribute to the design and development of Informatica master data management components of the solution in support of both member and provider master indexes.
  • Lead rules development for creation of a golden record for the most accurate domain content across member and provider possible.
  • Design and develop test plans for ETL unit testing and integration testing.
  • Provide training to the support team and customers.

Environment: Informatica Power Center 8.6/9.1, Oracle, SQL Server, PL/SQL, SQL Server Management Studio, SQL Developer, LINUX, Flat Files, Toad, Putty, Telnet, Tortoise SVN, Autosys and HP ALM Quality Center.

Confidential

ETL Consultant

Responsibilities:

  • Analyzing Business Requirements, Functional and Non Functional requirements to design and prepare technical specification documents.
  • Validating the source to target mappings, identify transformation rules to implement data conversion techniques.
  • Preparing ETL solution design documents describing the data load and data fresh strategy. Data Quality checks and Error Handling techniques.
  • Designing and creating the data warehouse and all related extraction, transformation and load of data functions.
  • Writing oracle queries, procedures and functions to store, retrieve and manipulate the data for enterprise applications.
  • Creating Informatica workflows and mappings to load history and incremental data using Change Data Capture mechanism.
  • Creating Unit Test Plans, Unit Test Cases, Minus Queries, Data Validation scripts and producing Unit Test Results.
  • Coding UNIX scripts to access, find and modify the delimited flat files to do SFTP/FTP to other files share locations.

Environment: Informatica 8.6, Oracle, UNIX, SQL Plus, Putty, Flat Files and HP ALM.

We'd love your feedback!