We provide IT Staff Augmentation Services!

Sr. Etl Data Engineer (lead) Resume

2.00/5 (Submit Your Rating)

Short Hills, NJ

SUMMARY:

  • 12+ years of IT experience in ETL, Data warehousing, Data Integration, Data Analytics, Data Engineering, Business Intelligence, Big Data, Data Migration and Data Modeling.
  • Experience in analysis, design, development, testing and implementation of Data Warehouse, Data Mart and Big Data based solutions.
  • Expertise in ETL process development and Data Integration solution architecture.
  • Deep understanding and experience working on ETL, Data Warehousing, Data Mart/Analytics and Big Data projects.
  • Extensive ETL development knowledge with tools like Informatica PowerCenter (Designer, Workflow Manager, Workflow Monitor and Server Manager) and Informatica PowerExchange. Developed data extraction, integration and loading using Informatica/ETL Workflows, Worklets, Mappings, Mapplets, Reusable transformations, scheduling and automation of the Workflows. Extensively developed ETL mappings using multiple ETL transformations including and not restricted to Expressions, Aggregator, Normalizer, Stored Procedure, Look Up, Update Strategy, Joiner, Filter, Sequence Generator, Sorter, Router in Informatica PowerCenter Designer, including performance tuning of mappings and using best coding principles to support scalable design.
  • Worked in different projects for various clients to build ETL and Data Warehousing based solution for customer products and analytical purposes, design and implementation of enterprise Data Warehouse/Data Mart for implementing end - to-end enterprise level data supply chain that interface with analytical and product-based applications.
  • Good knowledge in data warehousing concepts and multiple years of experience working with huge data driven and large data warehouse-based projects across different industries like Banking & Finance, Business Information & Data Analytics (Commercial Data, Credit Risk), Healthcare, Media & Entertainment.
  • Good understanding, hands on and usage of Dimensional Data modelling, ER Data model (Logical and Physical data model).
  • Good Knowledge in ETL development technology with utilization of tools like Informatica, DMExpress and Talend.
  • Experience in data conversion of various data sources like Oracle, SQL Server, Teradata, Mainframe, Flat Files (Fixed Width and Delimited), XML Files, VSAM files, Teradata Load/unload utilities M-load, Fast Load, BTEQ, T-pump, TPT, and Fast Export for loading and extracting data to and from a data warehouse/mart.
  • Strong development knowledge in Extraction, Transformation and Loading (ETL) data from various sources into Data Warehouse and Data Mart using Informatica Power Center (Repository Manager, Designer, Workflow Manager, Workflow Monitor, Metadata Manger) and Power Exchange on Oracle, Teradata, DB2 and SQL Server Databases.
  • Good understanding and knowledge in developing Slowly Changing Dimensions mapping (Type I, Type II & Type III) and Change data capture (CDC).
  • Developed Stored Procedures, Packages, Functions, Views and Triggers, Complex SQL queries in Oracle PL/SQL. This was developed for requirements of programs that needs to be executed on data warehouse / data mart and used along with requirements within ETL loads like Informatica.
  • Worked on Oracle Data Warehouse to Hadoop migration. Work performed in re-engineering project to migrate portion of ETL transformations written in DMExpress(Syncsort) and Informatica to Hadoop framework set up in AWS EC2 Instances and using data stored and handled with Hbase(Hfiles, Avro), Pig, Hive and Impala based on batch data processing requirements and downstream data usage.
  • Exposure and hands on experience with Data Integration and ETL using Hadoop HDFS framework implementation using MapReduce operations and data storage and SQL using Hbase/Hive/Impala and using transformations with Spark (Spark SQL, Python/Scala) for ETL and data processing.
  • Design and implementation of supplier portfolio data visualization dashboards and DUNS based detailed trade paydex and payment performance insight dashboards using Tableau.
  • Knowledge with on-premises enterprise data center and cloud-based services like Amazon Web Services (AWS).
  • Development experience in scripting language using UNIX shell scripts used for ETL workflows orchestration, data loads, data extraction, ad-hoc jobs and automation.
  • Development of Data reports and visualization knowledge using tools like Tableau
  • Database performance tuning and effective use of Database objects, SQL Trace, Explain Plan, different types of Optimizers, Hints, Indexes, Table Partitions, Sub Partitions, table stats, Materialized Views, Global Temporary tables, Capabilities of using Oracle Built-in Functions.
  • Worked with Informatica Data Quality toolkit, Analysis, data cleansing, data matching, data conversion, exception handling and monitoring capabilities of IDQ.
  • Code management and deployments working with Serena, Git, Stash, Jenkins and perform Continuous Integration.
  • Experience in Agile methodology and Scrum. Scrum Alliance Certified Scrum Master.
  • Very good working knowledge and skills in Microsoft Office suite (Word, PowerPoint, Excel, Visio).
  • Good knowledge and hands on development experience working in all phases of software development life cycle (SDLC) of Data warehousing projects, including and not restricting to requirement analysis, design, creation of mapping documents, specification writing, development, unit testing, integration testing, performance testing/tuning, implementation and post production.
  • Excellent interpersonal and communication skills with good experience in working with technology leaders, business owners, product teams, business analysts, QA testers and peer technical developers and team members.
  • Very good communication skills, strong decision making and organizational skills along with outstanding analytical and problem-solving ability to undertake challenging jobs. I will be able to work well independently and in a team by contributing to technology and business-related requirements and problems solving.

TECHNICAL SKILLS:

ETL: Informatica Power Center 9.6/9.1/8.6/8.5/8.1/7.1, Informatica PowerExchange, IDQ, Syncsort DMExpress 7.1/6.8, Informatica Data Quality (IDQ), Talend

Reporting: Tableau 10.x, SQL, Business Objects

Database/Data Warehousing: Oracle 10g/11g, Oracle Goldengate 11g, SQL-Server, Teradata 12.0, AWS Redshift, NoSQL (MongoDB, HBase)

Big Data: Hadoop, Hive, Sqoop, Spark(SQL), Impala, Cloudera, Oozie, Avro, Hue

Tools: and Utilities: Erwin, TOAD, SQL Developer, Teradata SQL Assistant, JIRA, Serena, Stash/Bitbucket, Jenkins, Rally, Quality Center(QC)/ALM, Ctrl-M, WinSCP, Putty, Word, PowerPoint, Excel, Visio, SharePoint, Skype, Microsoft Teams

Scripting, Programming Languages and Web Technologies: SQL, PL/SQL, T-SQL, Unix Shell Scripting, Python, HTML, XML, Perl, PHP

Cloud Computing: Amazon Web Services (AWS)

Operating Systems: Windows XP, Windows 10, Linux (Red Hat Enterprise Linux), MacOS

SDLC: Waterfall, Agile (Scrum, Kanban)

PROFESSIONAL EXPERIENCE:

Confidential,Short Hills, NJ

Sr. ETL Data Engineer (Lead)

Responsibilities:

  • Lead Extraction, Transformation and Loading (ETL) and Data Warehouse/Mart developer
  • Interacting with business and product owners to gather functional and technical requirements. Involved in extensive discussions with business and product members to understand and gather requirements.
  • Analyze business requirements and work on project design for ETL, Data Warehouse and Hadoop based data integration solution that will be used for both products and analytics. Technical documentation and working with business analysts.
  • Understand and analyze high level/low-level requirements, design documentation and developing technical documentation using source to target mapping specs, transformation rules, pseudo code, charts and process flow diagrams.
  • Develop ETL (Informatica) mappings, workflows, worklets, reusable objects based on program and project requirements.
  • Developed Stored Procedures, Packages, Functions, Views and Triggers, Complex SQL queries in Oracle PL/SQL. This was developed for requirements of programs that needs to be executed on data warehouse / data mart and used along with requirements within ETL loads like Informatica.
  • Perform technical code walkthrough, code review and technical discussions with technical team members.
  • Working with Informatica Data Quality (IDQ) for data cleansing, standardization and loading.
  • Developed Oracle DataMart for vending final aggregated data to customer facing credit risk insight products that connect using services layer.
  • Work on data modelling for the Global Trade Data Repository. This include modelling and designing tables in Staging, Warehouse and Data Mart layers specific to data requirements and relationships.
  • Prepare logic data model/diagrams, physical data model/diagrams, gather information about tables and attributes for data dictionary, prepare data model release checklist.
  • Development and implementation work performed in re-engineering project to migrate portion of ETL transformations written in DMExpress(Syncsort) and Informatica to Hadoop framework using Hbase, Hive and Impala based on the batch data processing requirements and downstream data usage.
  • Development and Implementation of data exports and imports using Sqoop to extract data from Oracle data warehouse to load data into Hadoop data sets with HDFS and MapReduce and loading data from Hadoop(Hive) into DataMart tables.
  • Develop Spark code using Spark (SQL), Python/Scala to join data from various base tables data present in HDFS and load modified data to and from Hive tables, fabricate, cleanse and transform data to create denormalized data to load intermediate data sets. The resultant data was loaded into specific tables in Impala and Redshift (Staging, Offline and Online tables). The data loaded into these tables can be used for dashboards and data analytics in web UI application called D&B Payment Performance Insights (PPI), Internal business operations reports and other products.
  • Developed Oozie job flows and coordinators for extraction and loading oracle data warehouse equivalent into Hadoop framework and data load from Hadoop(Hive) to Oracle.
  • Working with Cloudera Manager and Hue for interacting with Hadoop ecosystem and jobs.
  • Automation of ETL workflows including data cleansing, conversion, and transformations to load into Target data repositories.
  • Perform Data extraction and loading into Data Warehouse, Sqoop into Hadoop tables and integration with Product Mart and Analytics/Scoring applications.
  • Developed UNIX shell scripts for ETL workflows orchestration, data loads, data extraction, ad-hoc jobs and automation.
  • Developed various reports for Data quality, Data analytics and Error reports using SQL and Data visualization using Tableau.
  • Prepare Unit testing, system integration test cases; perform unit, system, regression, performance testing and help business users for user acceptance testing (UAT).
  • Working with Serena, BitBucket(Stash), Jenkins for code deployments, continuous integration, review and version control.
  • Worked with Oracle Golden Gate (OGG) for data replication between two application specific databases.

Confidential,Columbia, MD

ETL & SQL Developer (Lead)

Responsibilities:

  • Played a key role in the design and development phase of the anomaly reporting project.
  • Played an important role in understanding requirements of the client and transforming the information in an executable form.
  • Developed mapping specifications and design.
  • Coordinated with team members in analyzing the business requirements.
  • Played a major role in the Extract Transform and Load (ETL) team, which was responsible for coding the required business logic in Informatica mappings and developing executable workflows for the same.
  • Coordination with the team to render and gather information regularizing the development activities.
  • Created stored procedures, functions, scripts, and packages in Oracle PL/SQL as per requirements to load and perform various data operations and loads in Oracle database.
  • Worked with the team in understanding complex business requirements and communicating the technical aspects of the business to the team which helped the team members to work collectively for a better output.
  • Performance tuning and optimization achieved through the management of indices, table partitioning, and optimizing the SQL scripts.
  • Played major role in driving the testing phase of the project which involved unit testing, system integration testing and User acceptance testing (UAT).
  • Unit testing, performing defect analysis, performing code changes for fixing defects, integrating the application to perform integration testing and defect fixing during the same.

Confidential, Chicago, IL

ETL Developer

Responsibilities:

  • Developed several mappings using Informatica 7.1.3 that will be used to extract and load data into Operational Data Store (ODS)
  • Worked with performance tuning and optimization of mappings. Implemented some pushdown optimization techniques.
  • Perform Debugging & Tuning of mapping and sessions.
  • Developed business reports using Business Objects XI reporting tool.
  • Developed ad-hoc Teradata SQL reports for analyzing data load and errors.
  • Prepared BTEQ Scripts for loading of data into Operational Data Store (ODS).
  • Prepared Job Chains (Using Redwood Cronacle 6.0.2).
  • Created cron based scheduling system of all mappings, batches and ETL scripts in PERL and Unix Shell.
  • Used PERL and Shell scripts to pre-process data in UNIX
  • Developed automation batches and schedulers using Unix shell scripts.
  • Prepared Documents like Migration Documents, ETL Specifications, Unit Test Case, Performance Documents etc.
  • Develop unit test cases and perform testing.
  • Perform system testing, QA and integration testing.
  • Support & Coordinate with Migration team to Migrate Objects to QA environment.

Confidential,San Francisco, CA

ETL Data Warehouse Developer

Responsibilities:

  • Analyze and optimize the existing code.
  • Analysis, design, development, and implementation of data warehouse, ETL.
  • Used Informatica Source Analyzer, Mapping Designer and Warehouse Designer for Extraction, Transformation and Loading.
  • Extensively used Informatica Power Center and created mappings using Transformations and flagging the record using update strategy for populating the desired slowly changing dimension tables.
  • Worked on reporting requirements
  • Re-architected data flow for efficiency.
  • Simplified mappings by removing objects which are not needed.
  • Developed PL/SQL procedures for data extraction and loading.
  • Performance tuning of oracle procedures and queries for handling large volume of data.
  • Created complex mappings using Lookup, Filter, and Joiner and Router transformations for populating target table in an efficient manner.
  • Created UNIX shell scripts for calling Informatica workflows.
  • Used Parameterization for Workflows, Sessions.
  • Co-ordination and requirement gathering with onshore.
  • Performed Unit Tests & Reviews.
  • Perform system testing, QA and Integration testing
  • Played the lead role in rework team.
  • Worked on performance tuning of several complex mappings.
  • Issue Tracking, Defect Collection and Defect fixing.

We'd love your feedback!