We provide IT Staff Augmentation Services!

Data Engineer Resume

5.00/5 (Submit Your Rating)

Scottsdale, AZ

SUMMARY:

  • Eight plus years of experience in Analysis, Design, Development and Implementation as a Data Engineer.
  • Expert in providing ETL solutions for any type of business model.
  • Provided and constructed solutions for complex data issues.
  • Experience in development and design of various scalable systems using Hadoop technologies in various environments. Extensive experience in analyzing data using Hadoop Ecosystems including HDFS, MapReduce, Hive & PIG.
  • Experience in understanding the security requirements for Hadoop.
  • Extensive experience in working with Informatica Powercenter
  • Implemented Integration solutions for cloud platforms with Informatica Cloud.
  • Worked with Java based ETL tool, Talend.
  • Proficient in SQL, PL/SQL and Python coding.
  • Experience developing On - premise and Real Time processes.
  • Excellent understanding of best practices of Enterprise Data Warehouse and involved in Full life cycle development of Data Warehousing.
  • Expertise in DBMS concepts.
  • Involved in building Data Models and Dimensional Modeling with 3NF, Star and Snowflake schemas for OLAP and Operational data store (ODS) applications.
  • Skilled in designing and implementing ETL Architecture for cost effective and efficient environment.
  • Optimized and tuned ETL processes & SQL Queries for better performance.
  • Performed complex data analysis and provided critical reports to support various departments.
  • Work with Business Intelligence tools like Business Objects and Data Visualization tools like Tableau.
  • Extensive Shell/Python scripting experience for Scheduling and Process Automation.
  • Good exposure to Development, Testing, Implementation, Documentation and Production support.
  • Develop effective working relationships with client teams to understand and support requirements, develop tactical and strategic plans to implement technology solutions, and effectively manage client expectations.
  • An excellent team member with an ability to perform individually, good interpersonal relations, strong communication skills, hardworking and high level of motivation.

TECHNICAL SKILLS:

BigData Eco Systems: Hadoop, HDFS, MapReduce, Hive

Programming: Python

Data Warehousing: Informatica Power Center 9.x/8.x/7.x, Informatica Cloud, Talend Open studio & Integration suite

Applications: Salesforce, RightNow, Eloqua

Databases: Oracle (9i/10g/11g), SQL Server 2005

BI Tools: Business Objects XI, Tableau 9.1

Query Languages: SQL, PL/SQL, T-SQL

Scripting Languages: Unix, Python, Windows PowerShell

RDBMS Utility: Toad, SQL Plus, SQL Loader

Scheduling Tools: ESP Job Scheduler, Autosys, Windows scheduler

PROFESSIONAL EXPERIENCE:

Confidential, Scottsdale, AZ

Data Engineer

Responsibilities:

  • Analyze and cleanse raw data using HiveQL
  • Experience in data transformations using Map-Reduce, HIVE for different file formats.
  • Involved in converting Hive/SQL queries into transformations using Python
  • Performed complex joins on tables in hive with various optimization techniques
  • Created Hive tables as per requirements, internal or external tables defined with appropriate static and dynamic partitions, intended for efficiency
  • Worked extensively with HIVE DDLS and Hive Query language(HQLs)
  • Involved in loading data from edge node to HDFS using shell scripting.
  • Understand and manage Hadoop Log Files.
  • Manage Hadoop infrastructure with Cloudera Manager.
  • Created and maintained technical documentation for launching Hadoop cluster and for executing Hive queries.
  • Build Integration between applications primarily Salesforce.
  • Extensive work in Informatica Cloud.
  • Expertise in Informatica cloud apps Data Synchronization (ds), Data Replication (dr), Task Flows, Mapping configurations, Real Time apps like process designer and process developer.
  • Work extensively with flat files. Loading them into on-premise applications and retrieve data from applications to files.
  • Develop Informatica cloud real time processes (ICRT).
  • Work with WSDL, SOAP UI for APIs
  • Write SOQL queries, create test data in salesforce for informatica cloud mappings unit testing.
  • Prepare TDDs, Test Case documents after each process has been developed.
  • Identify and validate data between source and target applications.
  • Verify data consistency between systems.

Technologies Used: Bigdata ECO systems, Hadoop, HDFS, Hive, PIG, Cloudera, MapReduce, Python, Informatica Cloud Services, Salesforce, Unix scripts, FlatFiles, XML files

Confidential, Santa Clara, CA

Data Engineer

Responsibilities:

  • Built reporting data warehouse from ERP system using Order Management, Invoice & Service contracts modules.
  • Extensive work in Informatica Powercenter.
  • Acted as SME for Data Warehouse related processes.
  • Performed Data analysis for building Reporting Data Mart.
  • Worked with Reporting developers to oversee the implementation of report/universe designs.
  • Tuned performance of Informatica mappings and sessions for improving the process and making it efficient after eliminating bottlenecks.
  • Worked on complex SQL Queries, PL/SQL procedures and convert them to ETL tasks
  • Worked with PowerShell and UNIX scripts for file transfer, emailing and other file related tasks.
  • Worked with deployments from Dev to UAT, and then to Prod.
  • Worked with Informatica Cloud for data integration between Salesforce, RightNow, Eloqua, WebServices applications
  • Expertise in Informatica cloud apps Data Synchronization (ds), Data Replication (dr), Task Flows & Mapping configurations.
  • Worked on migration project which included migrating webmethods code to Informatica cloud.
  • Implemented Proof of concepts for SOAP & REST APIs
  • Built web services mappings and expose them as SOAP wsdl
  • Worked with Reporting developers to oversee the implementation of reports/dashboard designs in Tableau.
  • Assisted users in creating/modifying worksheets and data visualization dashboards in Tableau.
  • Tuned and performed optimization techniques for improving report/dashboard performance.
  • Assisted report developers with writing required logic and achieve desired goals.
  • Met End Users for gathering and analyzing the requirements.
  • Worked with Business users to identify root causes for any data gaps and developing corrective actions accordingly.
  • Created Ad hoc Oracle data reports for presenting and discussing the data issues with Business.
  • Performed gap analysis after reviewing requirements.
  • Identified data issues within DWH dimension and fact tables like missing keys, joins, etc.
  • Wrote SQL queries to identify and validate data inconsistencies in data warehouse against source system.
  • Validated reporting numbers between source and target systems.
  • Finding a technical solution and business logic for fixing any missing or incorrect data issues identified
  • Coordinating and providing technical details to reporting developers

Technologies Used: Informatica Power Center 9.5/9.1, Informatica Cloud, Oracle 10g/11g, SQL Server 2005, Tableau 9.1, Salesforce, RightNow, Eloqua, Web Methods, PowerShell, Unix

Confidential, San Mateo, CA

Data Engineer

Responsibilities:

  • Implemented reporting Data Warehouse with online transaction system data.
  • Developed and maintained data warehouse for PSN project.
  • Provided reports and publications to Third Parties for Royalty payments.
  • Managed user account, groups and workspace creation for different users in Powercenter.
  • Wrote complex UNIX/windows scripts for file transfers, emailing tasks from FTP/SFTP.
  • Worked with PL/SQL procedures and used them in Stored Procedure Transformations.
  • Extensively worked on oracle and SQL server. Wrote complex sql queries to query ERP system for data analysis purpose
  • Worked on most critical Finance projects and had been the go to person for any data related issues for team members.
  • Migrated ETL code from Talend to Informatica. Involved in development, testing and post production for the entire migration project.
  • Documented the code.
  • Tuned ETL jobs in the new environment after fully understanding the existing code.
  • Maintained Talend admin console and provided quick assistance on production jobs.
  • Involve in designing Business Objects universes and creating reports.
  • Built adhoc reports using stand-alone tables.
  • Involved in creating and modifying new and existing Web Intelligence reports.
  • Created Publications which split into various reports based on specific vendor.
  • Wrote Custom SQL for some complex reports.
  • Worked with business partners internal and external during requirement gathering.
  • Worked closely with Business Analyst and report developers in writing the source to target specifications for Data warehouse tables based on the business requirement needs.
  • Exported data into excel for business meetings which made the discussions easier while looking at the data.
  • Performed analysis after requirements gathering and walked team through major impacts.
  • Provided and debugged crucial reports for finance teams during month end period.
  • Addressed issue reported by Business Users in standard reports by identifying the root cause.
  • Get the reporting issues resolved by identifying whether it is report related issue or source related issue.
  • Creating Ad hoc reports as per users needs.
  • Investigating and Analysing any discrepancy found in data and then resolving it.

Technologies Used: Informatica Power Center 9.1/9.0, Talend 4.x & Integration suite, Business Objects XI, Oracle 10g/11g, Oracle ERP, EDI, SQL Server 2005, UNIX, Windows Scripting, JIRA

Confidential, San Diego, CA

ETL Developer

Responsibilities:

  • Gathered business requirements and prepared technical design documents, target to source mapping document, mapping specification document.
  • Extensively worked on Informatica Powercenter.
  • Parsed complex files through Informatica Data Transformations and loaded it to Database.
  • Optimized query performance by oracle hints, forcing indexes, working with constraint based loading and few other approaches.
  • Extensively worked on UNIX Shell Scripting for splitting group of files to various small files and file transfer automation.
  • Worked with Autosys scheduler for scheduling different processes.
  • Performed basic and unit testing.
  • Assisted in UAT Testing and provided necessary reports to the business users.

Technologies Used: Informatica Power Center 8.6, Oracle 10g/11g, UNIX Shell Scripting, Autosys

Confidential, Irving, TX

ETL/Data Warehouse Developer

Responsibilities:

  • Gathered requirements from Business and documented for project development.
  • Coordinated design reviews, ETL code reviews with teammates.
  • Developed mappings using Informatica to load data from sources such as Relational tables, Sequential files into the target system.
  • Extensively worked with Informatica transformations.
  • Created datamaps in Informatica to extract data from Sequential files.
  • Extensively worked on UNIX Shell Scripting for file transfer and error logging.
  • Scheduled processes in ESP Job Scheduler.
  • Performed Unit, Integration and System testing of various jobs.

Technologies Used: Informatica Power Center 8.6, Oracle 10g, SQL Server 2005, UNIX Shell Scripting, ESP job scheduler

Confidential, St. Louis, MO

Data Analyst

Responsibilities:

  • Involved in understanding the legacy applications & data relationships.
  • Attended user design sessions, studied user requirements, completed detail design analysis and wrote design specs.
  • Interacted with key users and assisted them with various data issues, understood data needs and assisted them with Data analysis.
  • Prepared and maintained documentation for on-going projects.
  • Worked with Informatica PowerCenter for data processing and loading files.

Technologies Used: Informatica PowerCenter 7.x, Oracle 9i, SQL, Toad, SQL Plus

We'd love your feedback!