Data Engineer Resume
Glendale, CA
SUMMARY
- 10+ years of IT experience in Data Engineering Cloud Migration projects.
- Experience in Snowflake/Informatica Cloud {IICS}/Snap Logic/AWS/Azure/Redshift/S3/Json Files/Python and EC2 Instance Experience using ETL tool Informatica Power Center.
- Worked in various domains which includes Hospitality, Insurance, healthcare, banking & environmental sciences.
- Experience in providing the ETL solutions, Data Analysis and Data profiling.
- Strong Experience in Snowflake and experience working in different sources which includes Redshift, s3, xml, relational databases, Salesforce, Oracle Apps, DB2, Sql Server, Postgres, Netezza & Teradata.
- Good exposure in Azure Databricks cloud platform.
- Experience working in Data Warehousing Dimension Tables, Fact tables, slowly changing dimensions, DataMart’s and dimensional modeling schemas.
- Extensively worked with data warehousing, ETL methodologies, technologies, Informatica Power Center.
- Experience in Informatica MDM.
TECHNICAL SKILLS
ETL Tools: IICS Informatica Cloud, Snap Logic/Informatica MDM, AWS Glue
Cloud Technology: AWS EC2,S3, Google Cloud Platform (GCP)/Microsoft Azure Azure Databricks /Salesforce.com, Salesforce CRM, SOQL and Apex Data Loader.
Operating Systems: Windows 10/7/2010/XP/2000/NT, HP - Unix, Linux
Schedulers: Control-M, AutoSys, Cybermation ESP
Languages: Sql, PL/Sql, T Sql, Python, PySPark
Databases: Snowflake, Google Big Query, Oracle 12c,11g,10g,9i,8i, SQL Server 2005, DB2, Netezza, Teradata.
PROFESSIONAL EXPERIENCE
Confidential, Glendale CA
Data Engineer
Responsibilities:
- Worked in Agile methodology and experience all agile ceremonies. Worked on understanding the project requirements and create the Integration solution design. Worked in Snowflake Database, created tables, views, temporary tables and reporting tables. Created etl mappings in IICS Informatica Intelligent Cloud Services.
- Created Azure Databricks pipelines to fetch the data from Snowflake Raw Schema to Modeled Schema.
- Created Stages, worked on time travel scripts, zero copy cloning and materialized views in snowflake.
- Build complex ETL processes that transform data using Azure Databricks.
- Work on a data factory pipeline where we are ingesting data from Amazon S3 to Azure for processing the ingested data using a Notebook running in Azure Databricks
- Created and transform (clean, sort, merge, join, etc.) the ingested data in Azure Databricks as aNotebook
- Worked on AWS cloud services: EC2, S3. Data ingestion using REST APIs into Snowflake/Azure and its orchestration. Technical/Functional Skills on Azure Data Lake ADLS, Experience in Azure Data Bricks, ADF, ADL, Hive, Python, Databricks, Spark.
- Translate business logic to SQL queries. Created reports for business based on the fact and dimension data.
- Written Sql queries and built Reconciliation reports on Claims, Estimates, Notifications and Customers data to help business validate the data loss between source to target points. Involved in snowflake tables data modeling and prepared the data analysis
Environment: Azure, Snowflake, JSon Files, Informatica IICS, Linux, AWS, ADLS, Oracle, MS Office 365, Python 2.7.5
Confidential, Chevychase MD
Data Engineer
Responsibilities:
- Worked in Agile methodology and experience all agile ceremonies.
- Worked on understanding the project requirements and create the Integration solution design.
- Worked in Columnar AWS Redshift and NoSQL Postgres Databases.
- Import data into Google Big query from the data files and query the data.
- Coordinated with offshore team prepared design documents, reviewed the code, deployed and provided QA/UAT Support.
- Knowledge of AWS cloud services: EC2, S3, Glue. Build Semantic Reporting Layer in Google Big Query.
- Dataflow -Big Query -GCS, Experience with Hadoop ecosystem, Experience with google cloud (GCP) Data pipeline migration & optimizationand Understandings on infrastructure automation and Experience working with Scala and Pyspark.
- Translate business logic to SQL queries. Created reports for business based on the fact and dimension data.
- Written Sql queries and built Reconciliation reports on Brands, products, Consumer and Customers data to help business validate the data loss between source to target points. Data ingestion using REST API into Snowflake/Azure by creating the orchestration pipelines.
- Involved in redshift tables data modeling and prepared the data analysis which helps to identify the right distribution keys and sort keys in Redshift data base based on the data. CICD Deployment process and continuous Integration.
- Used Anaconda as python interpreter, and automate the jobs using pyspark applications.
- Understand the existing & new SFDC data model & prepare a feasible ETL solution for the data migration into SFDC.
- Perform data designing, model & created MDM landing and base objects, mappings, executed stage and load jobs.
- Designed developed, deployed and migrated the ETL code and supported the data load jobs in Dev/QA/Prod Environments.
Environment: AWS, Redshift, Informatica Powercenter, Linux, Oracle, Google Big Query, DB2, Anaconda, PySpark, MS Office 2013, Oracle EBS, Salesforce
Confidential, Bridgewater NJ
ETL/MDM Consultant
Responsibilities:
- Experience working on the Production MDM Jobs.
- Created Batch groups executed them in sequence and in parallel for each source system.
- Involved in designing MDM data model and created base objects, mappings, defined trust setting for sources.
- Design and configuration of landing tables, staging tables, base objects, hierarchies, foreign-key relationships, lookups, query groups, queries/custom queries and packages in MDM.
- Created users, assign roles and privileges in the Hub Console.
- Implemented Stage Jobs, Load Jobs, Match and Merge Jobs using the Batch Viewer and Automation Processes.
- Involved in Migration of objects from Dev to QA and QA to Prod Informatica MDM 9.7.1.
- Imported and exported the IDD, ORS, Queries from Hub Console and IDD and Install Siperian ear files.
- Experience in Informatica MDM HUB development support & technical knowledge of Informatica MDM products
Environment: MS Office 2010, Informatica MDM, Informatica IICS, IDD, 9.7.1, CICD, Linux 2.6, DB2 v9.7.0.9, Informatica PowerCenter 9.6.1
Confidential, Dallas Tx
Informatica ETL Lead
Responsibilities:
- Participate in requirement gathering, validation, design, development, and maintenance of BI solutions
- Tuning Informatica PowerCenter Mappings and Sessions for optimum performance.
- Design and configuration of landing tables, staging tables, base objects, hierarchies, foreign-key relationships, lookups, query groups, queries/custom queries and packages in MDM.
- Interacted with business users in configuring Match setup and tool proposition for Data quality Address doctor.
- Implemented Stage Jobs, Load Jobs, Match and Merge Jobs using the Batch Viewer and Automation Processes.
- Experience in Informatica MDM HUB development support & detailed technical knowledge of Informatica MDM products
- Created ETL Mappings to load Oracle data to Salesforce.com, created Salesforce objects and worked in SOQL.
Environment: MS Office 365, Oracle 12C Enterprise Edition, Informatica Cloud, Informatica PowerCenter 9.6.1, UNIX, Salesforce.com, Apex Data Loader 36.0, Informatica IDQ
Confidential, Boston MA
Lead ETL Engineer
Responsibilities:
- Created ETL Mappings to load Oracle data to Salesforce.com, created Salesforce objects and worked in SOQL.
- Responsible for understanding, articulating, documenting and presenting DB requirements on the assigned projects.
- Involved in the development of Informatica PowerCenter mappings.
- Develop the oracle pl/sql store procedures which does the core ETL Operation with in oracle tables.
- Tuned Informatica power center Mappings and Sessions for optimum performance.
- Supporting requirement analysis, system analysis, design, development, testing, implementation
- Tuning Informatica Powercenter Mappings and Sessions for optimum performance.
- Monitor the jobs running using informatica powercenter workflow monitor tool ensure the completion of the deliverables.
Environment: MS Office 2010, BMC Control M 7, Informatica Powercenter 9, Informatica Cloud, Informatica Power Exchange, Oracle 11g, UNIX, Salesforce.com, Salesforce Dataloader 34.0
Confidential, Boston MA
Senior ETL Consultant
Responsibilities:
- Responsible for understanding, articulating, documenting and presenting DB requirements (both business and technical) on the assigned projects.
- Extracted data from SQL Server to Teradata using Informatica.
- Create the views, sequences, temp tables, ref tables and external tables while creating stored procedures.
- Worked with Teradata BTEQ, Fastexport, TPT, Fastload, Multiload (mload) utilities.
- Developed Informatica mappings and also tuned for testing the functionalities of the PWX for Netezza.
- Tuned Mappings and Sessions to optimize and enhance performance.
- Responsible for understanding, articulating, documenting and presenting DB requirements on the assigned projects
- Extracted data from SQL Server, Oracle and Teradata to Netezza using Informatica.
- Monitor the production ETL jobs fix and support the issues. Re execute the batch jobs after the fix.
- Extensively worked on Netezza Database for the best performance on the data loads.
- Interact with DB, Applications analysts and Business users for requirements gathering and validation.
- Involved in the Conduct code review meetings of the team members.
- Provided impact analysis for the changes implemented and effect on the database and systems.
Environment: Informatica Powercenter9, Oracle 10g, Netezza 7, Teradata, MPP, UNIX AIX 5, Microsoft XP Pro, Autosys
Confidential
ETL Developer
Responsibilities:
- Responsible for ETL Informatica powercenter mapping development according to the STTM Mappings.
- Developed mapping to load the data in slowly changing dimension.
- Worked extensively with SQL*Plus, TOAD for user defined queries, stored pro’s, views, tables, indexes, etc.
- Worked on extracting data from Oracle, SQL Server database transform and loading into Data Warehouse.
Environment: Informatica Powercenter 7.1.2, Oracle 9i, SQL Server 2000, UNIX, Linux, Windows.
