We provide IT Staff Augmentation Services!

Data Integration Specialist Resume

0/5 (Submit Your Rating)

Pittsburgh, PA

SUMMARY

  • 9 years of Intelligent Data Integration and ETL experience in application development with the large Enterprise Data Warehouse and Business Intelligence systems.
  • Strong Knowledge on contemporary Data Warehousing trends - Bill Inman, Ralph Kimball methodologies, Star Schema, Snowflake schema, ODS, EDW, DM, OLAP Dimensions and Facts.
  • Analyze, validate and refine Business and Functional requirements to implement the strategic Data Warehouse Life Cycle for efficient Business Intelligence trends.
  • Develop and make enhancements to new/existing dimensional models in preparing scalable, reliable and optimized physical, logical and conceptual data models.
  • Work with Application Architects in defining project specific Blue Prints and Architectural diagrams.
  • Prepare Source to Target Data Mappings, ETL Design and technical specification documents in compliance with the data governance and best standards and review with the Enterprise Architects and ETL Admins.
  • Develop PLSQL procedures, functions and packages that store, retrieve and manipulate data for data analytics. Proficient in using CLOB, BLOB and LONG datatype based blocks for Stored Procedures.
  • Develop reusable code snippets, automated data movement reconciliation and Batch Framework techniques.
  • Implement Error Logging and Alert/Abort process for any data flow issues and enable restart ability/recover techniques through Informatica Power Center tool.
  • Perform Data Validation and Reconcilement checks for each data movement process and log the results for auditing purpose. Deploy the ETL/Oracle/Linux code through DevOps tools like Git, uDeploy and Jenkins.
  • Implement Data Retention techniques for auditing purpose and automate the Purge mechanism for very old/unused data extracts.
  • Proficient in leveraging latest databases Oracle 12c, Teradata and Netezza in achieving better performance through Partitioning, Optimizers and MPP (Massively Parallel Processing) architecture.
  • Perform troubleshooting on production issues, identify application bottlenecks and opportunities to resolve the errors and optimize the performance.
  • Work with ETL admins in setting up Operating System Profiles (OSP), Red Hat Linux NAS/SAN locations, Domain, Node and Repository Configurations, Version Control and Deployments to higher environments.
  • Work with DBA in sizing data and index table spaces for database creation, creating and granting system privileges/roles, deploying DDL/DML scripts to higher environments.
  • Strong knowledge on Cloud Integration, Big Data, Hadoop File System, Financial and Insurance markets.

TECHNICAL SKILLS

Data Warehouse: Ralph Kimball and Bill Inman Methodologies, SCD Type 1 or Type 2, Star Schema, Snow Flake Schema, Dimensions, Facts, OLTP, ODS, EDW, DM and OLAP.

Data Model: Dimensional Model and ER Models.

Databases: Teradata, Netezza and Oracle 12c.

Files: Fixed Width Flat Files, Delimited Flat Files and XML.

Languages: TSQL, NZSQL, SQL, PLSQL and Shell Script

TOOLS: Data Integration: Informatica Power Center Tools (10.x or 9.x)

Data Model: Erwin and Visio

Database: Teradata Studio, AQT, TOAD, SQL Developer and SQL Plus.

FTP: Putty, Ultra Edit and WinSCP

DevOps Tools: GIT, uDeploy, Artifactory, Jenkins, Tortoise SVN and CA Software Change Manager

Schedulers: CA7, CICS Explorer, ESP (CA WA Workstation) and Autosys.

Testing: HP ALM Quality Center

Incident Management: Service Now, Oracle Identity Manager (OIM)

PROFESSIONAL EXPERIENCE

Confidential, Pittsburgh, PA

Data Integration Specialist

Responsibilities:

  • Analyzing, Validating and Refining Business Requirements, Functional and Non-Functional requirements, defining development scope and technical specification documents.
  • Performing deep dive analysis on upstream systems to identify the Key Performance Indicators (KPI), transformation rules, data load strategy and data gaps in order to accomplish project data requirements.
  • Developing or enhancing dimensional models in compliance with the Data Governance and Stewardship standards and fit it into Enterprise Architecture.
  • Preparing Source to Target Mappings (STM) for each data movement process from heterogeneous sources to staging and then to Data Warehouse and Data Marts.
  • Designing each ETL process with complete data load and refresh strategy, data reconcilement and validation techniques, Error Logging and Restart ability/Recover Strategy, Data Retention and Archival/Purge mechanism.
  • Designing Batch Framework with complete job statistics that include Job Name, status of the job, load frequency, source/target data volume, execution start and end dates, dependency and error details, etc.
  • Developing Informatica Mappings, Sessions and Workflows for each data movement process that include Source to Work (Truncate and Load), Work to Stage (Insert/Persistent) and Work to DIM/FACT (Change Data Capture).
  • Coding PLSQL procedures, triggers, packages and DB components (Stored Procedures) which can be leveraged in multiple applications and avoid repetitive code.
  • Preparing Database Scripts (DDL, DML, DCL and TCL), Linux/Shell scripts and Scheduling Batch Scripts.
  • Conducting Code Reviews with ETL and DB Admins in order to comply with the Industry Best Practices and Development Standards.
  • Performing Data Quality Checks and Data profiling to cleanse the source data using Informatica IDQ tool.
  • Implement version control for ETL and DB components through GIT and Artifactory tools and deploy them using IBM uDeploy and Jenkins tools.
  • Perform troubleshooting on long running jobs and identify the opportunities to optimize the performance.
  • Install third party (KPMG) database gClas (global Credit loss accounting system), map the gClas to DW and ingest the file from Data Warehouse to forecast the expected credit losses.
  • Creating Journal Entries after successful reconciliation of balances, accounts from the General Ledger.
  • Scheduling the Informatica Workflows or jobs using CA7 scheduler based on operational calendar.
  • Implementing Data Masking techniques to protect Personally Identifiable Information (PII) from the users.
  • Good knowledge on Comprehensive Capital Analysis and Review (CCAR) models and Sarbanes-Oxley (SOX) Act requirements.

Environment: Informatica Power Center Tools 10.1, Teradata, Oracle 12c, SQL, LINUX, Flat Files, Teradata Studio, Toad, Putty, GIT, Artifactory, uDeploy, CA7 scheduler, Oracle Identity Manager, HP ALM Quality Center, Ultra Edit, WinSCP and Service now.

Confidential, Columbus, OH

Sr. ETL Developer

Responsibilities:

  • Analyze business requirement, define development scope and develop technical specification documents.
  • Code, test, modify, debug, document, and implement ETL workflows using Informatica Power Center tools.
  • Solid experience with data modeling and database design for operational as well as analytical systems.
  • Develop scripts to automate the execution of ETL Workflows/mappings using shell scripts under UNIX environment.
  • Adhere to IT Control Policies throughout design, development and testing and in corporate Corporate Architectural Standards into application design specifications.
  • Participate in code & design reviews and ensure that all solutions are aligned to pre-defined architectural specifications; identify/troubleshoot application code-related issues; and review and provide feedback to the final user documentation.
  • Develop Batch Framework through Perl scripts which set Informatica Environment, Configure Database Connections, Trigger Informatica workflows, Update Job Status tables and create daily log files.
  • Extensive experience in writing PLSQL blocks including Stored Procedures, Functions, Triggers and Packages with Cursors, CLOB, BLOB and LONG datatypes. Developed error handling and data retention procedures.
  • Developed Change Data Capture flows for incremental loads, PLSQL blocks that perform gather stats, data archiving and purging during DB maintenance window.
  • Version and revision control practices and procedures for all ETL components including Informatica Workflows, DB scripts, Parameter files and configuration files thru Tortoise SVN and CA Software Change Manager.
  • Develop ESP Jobs, Events and applications through CA WA Workstation and CICS Explorer tools.
  • Develop Data Quality scripts that will validate the source and target systems after loading/generating the data for downstream systems.

Environment: Informatica Power Center Tools 9.6.1 Hot Fix, Oracle, PL/SQL, SQL, LINUX, Flat Files, XML, Perl, Toad, Putty, Tortoise SVN, CA Software Change Manager, CICS Explorer, ESP (CA WA Workstation), HP ALM Quality Center, Ultra Edit, WinSCP and Service now.

Confidential, Columbus, OH

Sr. ETL Developer

Responsibilities:

  • Understand Business, Functional Requirements and Data Mappings for ETL design and implementation.
  • Develop ETL processes using Informatica Power Center tools to implement logical and physical dataflow models according to functional requirements and technical designs.
  • Develop slowly changing Dimension (SCD Type 1, Type 2) to load Dimension Tables and implemented Change Data Capture (CDC) mechanism for Incremental Load.
  • Develop SQL/PLSQL queries and ETL (extract, transform, and load data) that is consistent and integrated with existing ETL processes.
  • Code, test, and implement Informatica mappings and scripts to build and maintain automated ETL processes.
  • Fix high priority production issues and support daily/weekly/monthly running jobs by partnering with other IT areas in resolving issues and improving processes.
  • Perform Data Quality validations for each data movement flow that comprises of critical balances and counts.
  • Provide resolution to an extensive range of complicated ETL-related problems, proactively and as issues surface.
  • Coordinate and develop process automation/job scheduling, perform testing and defect analysis.
  • Work with technical leads, architects, subject matter experts, developers, technical support groups and test teams to deliver value to the business.
  • Develops Data Quality rules and mappings based on the requirements and technical design.
  • Participate in design sessions with ETL developers, report developers, DBAs and Business Analysts.
  • Remain current on new ETL techniques and methodologies and communicate trends and opportunities to management and other developers as needed.
  • Oversees development of project plans, including requirements and level of effort (LOE).
  • Ensures communication of project status to leadership and issues within team.

Environment: Informatica Power Center 9.6/9.5 Hot Fix, Erwin, Teradata, DB2, Netezza, Oracle, PL/SQL, SQL, LINUX, Flat Files, XML, AQT, Toad, Putty, Tortoise SVN, CA SCM, CICS Explorer, ESP (CA WA Workstation), HP ALM Quality Center, Ultra Edit, WinSCP and Service now.

Confidential, Hartford, CT

ETL Developer

Responsibilities:

  • Interact with business analysts and end users to understand requirements and convert business specifications to technical artifacts and determine most efficient design solution across multiple integration projects
  • Adhere to architecture standards, risk management and security policies, leading presentations in architecture review, strategic technology directions, best practice development (eg, estimating models), mentoring less experienced team members, code reviews
  • Design, analyze, develop, code, test, debug and document programming to satisfy business requirements
  • Develop Change Data Capture flows using Informatica, Oracle, Unix (Korn Shell scripts), PL/SQL blocks.
  • Thoroughly perform unit testing ensuring minimal code defects out of build phase into system test phase
  • Good communication skills and solid knowledge on software delivery lifecycle in large Warehouse environment
  • Implement Data Quality techniques for each incremental load that captures changes through Change Data Capture mechanism from the upstream systems.
  • Extensive Hands on experience designing and developing all Data Integration flows using Informatica.
  • Well versed in the following data domains: Master Data, Operational Data, Analytical Data, Unstructured Data
  • Complete code reviews for ETLs and related components, and complete documentation of issues identified and action items.
  • Correct testing defects and support all testing, including but not limited to: Development Integration Testing, System Testing, User Acceptance Testing, End-to-End Testing, and Performance Testing.
  • Provide resolution to an extensive range of complicated ETL-related problems, proactively and as issues surface.
  • Work with initiative leads, business areas, technical leads, architects, subject matter experts, developers, technical support groups and test teams to deliver value to the business.

Environment: Informatica Power Center 9.5.1, Teradata, Oracle, PL/SQL, SQL, SQL Loader, LINUX, Flat Files, Toad, Putty, Tortoise SVN, Autosys, HP ALM Quality Center and Ultra Edit.

Confidential, Hartford, CT

ETL Developer

Responsibilities:

  • Create data mapping and workflow using Informatica PowerCenter to extract, transform and load data into the target reporting environment.
  • Conform to Informatica workflows and mapping standards so that development is done in a consistent manner across all development resources (e.g., no hard coding, leverage of mapplet / reuse concept).
  • Develop SCD Type 1, Type 2 using Change Data Capture to maintain history of transactions for Incremental .
  • Tune and optimize ETL jobs for performance and throughput. Interact with data modelers and DBAs on data architecture optimization for ETL performance.
  • Provide development support, including walk through approvals, to other ETL application resources to insure that standards are followed and optimized workflow is implemented.
  • Advocate and enforce secure coding principles, as defined coding policies, to insure we are in compliance with all defined standards.
  • Assist in documenting and maintaining operational procedures in the form of an operations manual for all ETL workflows.
  • Writing Data Quality validation procedures to enable data completeness strategy for accurate information for BI
  • Assist with troubleshooting efforts for ETL job issues related to pre-production and production.
  • Leverage Data Quality as a mandatory process step for profiling data prior to downstream transformation.
  • Contribute to the design and development of Informatica master data management components of the solution in support of both member and provider master indexes.
  • Lead rules development for creation of a golden record for the most accurate domain content across member and provider possible.
  • Design and develop test plans for ETL unit testing and integration testing.
  • Provide training to the support team and customers.

Environment: Informatica Power Center 8.6/9.1, Oracle, SQL Server, PL/SQL, SQL Server Management Studio, SQL Developer, LINUX, Flat Files, Toad, Putty, Telnet, Tortoise SVN, Autosys and HP ALM Quality Center.

Confidential

ETL Consultant

Responsibilities:

  • Analyzing Business Requirements, Functional and Non Functional requirements to design and prepare technical specification documents.
  • Validating the source to target mappings, identify transformation rules to implement data conversion techniques.
  • Preparing ETL solution design documents describing the data load and data fresh strategy. Data Quality checks and Error Handling techniques.
  • Designing and creating the data warehouse and all related extraction, transformation and load of data functions.
  • Writing oracle queries, procedures and functions to store, retrieve and manipulate the data for enterprise applications.
  • Creating Informatica workflows and mappings to load history and incremental data using Change Data Capture mechanism.
  • Creating Unit Test Plans, Unit Test Cases, Minus Queries, Data Validation scripts and producing Unit Test Results.
  • Coding UNIX scripts to access, find and modify the delimited flat files to do SFTP/FTP to other files share locations.

Environment: Informatica 8.6, Oracle, UNIX, SQL Plus, Putty, Flat Files and HP ALM.

We'd love your feedback!