Data Engineer Resume
Scottsdale, AZ
SUMMARY:
- Eight plus years of experience in Analysis, Design, Development and Implementation as a Data Engineer.
- Expert in providing ETL solutions for any type of business model.
- Provided and constructed solutions for complex data issues.
- Experience in development and design of various scalable systems using Hadoop technologies in various environments. Extensive experience in analyzing data using Hadoop Ecosystems including HDFS, MapReduce, Hive & PIG.
- Experience in understanding the security requirements for Hadoop.
- Extensive experience in working with Informatica Powercenter
- Implemented Integration solutions for cloud platforms with Informatica Cloud.
- Worked with Java based ETL tool, Talend.
- Proficient in SQL, PL/SQL and Python coding.
- Experience developing On - premise and Real Time processes.
- Excellent understanding of best practices of Enterprise Data Warehouse and involved in Full life cycle development of Data Warehousing.
- Expertise in DBMS concepts.
- Involved in building Data Models and Dimensional Modeling with 3NF, Star and Snowflake schemas for OLAP and Operational data store (ODS) applications.
- Skilled in designing and implementing ETL Architecture for cost effective and efficient environment.
- Optimized and tuned ETL processes & SQL Queries for better performance.
- Performed complex data analysis and provided critical reports to support various departments.
- Work with Business Intelligence tools like Business Objects and Data Visualization tools like Tableau.
- Extensive Shell/Python scripting experience for Scheduling and Process Automation.
- Good exposure to Development, Testing, Implementation, Documentation and Production support.
- Develop effective working relationships with client teams to understand and support requirements, develop tactical and strategic plans to implement technology solutions, and effectively manage client expectations.
- An excellent team member with an ability to perform individually, good interpersonal relations, strong communication skills, hardworking and high level of motivation.
TECHNICAL SKILLS:
BigData Eco Systems: Hadoop, HDFS, MapReduce, Hive
Programming: Python
Data Warehousing: Informatica Power Center 9.x/8.x/7.x, Informatica Cloud, Talend Open studio & Integration suite
Applications: Salesforce, RightNow, Eloqua
Databases: Oracle (9i/10g/11g), SQL Server 2005
BI Tools: Business Objects XI, Tableau 9.1
Query Languages: SQL, PL/SQL, T-SQL
Scripting Languages: Unix, Python, Windows PowerShell
RDBMS Utility: Toad, SQL Plus, SQL Loader
Scheduling Tools: ESP Job Scheduler, Autosys, Windows scheduler
PROFESSIONAL EXPERIENCE:
Confidential, Scottsdale, AZ
Data Engineer
Responsibilities:
- Analyze and cleanse raw data using HiveQL
- Experience in data transformations using Map-Reduce, HIVE for different file formats.
- Involved in converting Hive/SQL queries into transformations using Python
- Performed complex joins on tables in hive with various optimization techniques
- Created Hive tables as per requirements, internal or external tables defined with appropriate static and dynamic partitions, intended for efficiency
- Worked extensively with HIVE DDLS and Hive Query language(HQLs)
- Involved in loading data from edge node to HDFS using shell scripting.
- Understand and manage Hadoop Log Files.
- Manage Hadoop infrastructure with Cloudera Manager.
- Created and maintained technical documentation for launching Hadoop cluster and for executing Hive queries.
- Build Integration between applications primarily Salesforce.
- Extensive work in Informatica Cloud.
- Expertise in Informatica cloud apps Data Synchronization (ds), Data Replication (dr), Task Flows, Mapping configurations, Real Time apps like process designer and process developer.
- Work extensively with flat files. Loading them into on-premise applications and retrieve data from applications to files.
- Develop Informatica cloud real time processes (ICRT).
- Work with WSDL, SOAP UI for APIs
- Write SOQL queries, create test data in salesforce for informatica cloud mappings unit testing.
- Prepare TDDs, Test Case documents after each process has been developed.
- Identify and validate data between source and target applications.
- Verify data consistency between systems.
Technologies Used: Bigdata ECO systems, Hadoop, HDFS, Hive, PIG, Cloudera, MapReduce, Python, Informatica Cloud Services, Salesforce, Unix scripts, FlatFiles, XML files
Confidential, Santa Clara, CA
Data Engineer
Responsibilities:
- Built reporting data warehouse from ERP system using Order Management, Invoice & Service contracts modules.
- Extensive work in Informatica Powercenter.
- Acted as SME for Data Warehouse related processes.
- Performed Data analysis for building Reporting Data Mart.
- Worked with Reporting developers to oversee the implementation of report/universe designs.
- Tuned performance of Informatica mappings and sessions for improving the process and making it efficient after eliminating bottlenecks.
- Worked on complex SQL Queries, PL/SQL procedures and convert them to ETL tasks
- Worked with PowerShell and UNIX scripts for file transfer, emailing and other file related tasks.
- Worked with deployments from Dev to UAT, and then to Prod.
- Worked with Informatica Cloud for data integration between Salesforce, RightNow, Eloqua, WebServices applications
- Expertise in Informatica cloud apps Data Synchronization (ds), Data Replication (dr), Task Flows & Mapping configurations.
- Worked on migration project which included migrating webmethods code to Informatica cloud.
- Implemented Proof of concepts for SOAP & REST APIs
- Built web services mappings and expose them as SOAP wsdl
- Worked with Reporting developers to oversee the implementation of reports/dashboard designs in Tableau.
- Assisted users in creating/modifying worksheets and data visualization dashboards in Tableau.
- Tuned and performed optimization techniques for improving report/dashboard performance.
- Assisted report developers with writing required logic and achieve desired goals.
- Met End Users for gathering and analyzing the requirements.
- Worked with Business users to identify root causes for any data gaps and developing corrective actions accordingly.
- Created Ad hoc Oracle data reports for presenting and discussing the data issues with Business.
- Performed gap analysis after reviewing requirements.
- Identified data issues within DWH dimension and fact tables like missing keys, joins, etc.
- Wrote SQL queries to identify and validate data inconsistencies in data warehouse against source system.
- Validated reporting numbers between source and target systems.
- Finding a technical solution and business logic for fixing any missing or incorrect data issues identified
- Coordinating and providing technical details to reporting developers
Technologies Used: Informatica Power Center 9.5/9.1, Informatica Cloud, Oracle 10g/11g, SQL Server 2005, Tableau 9.1, Salesforce, RightNow, Eloqua, Web Methods, PowerShell, Unix
Confidential, San Mateo, CA
Data Engineer
Responsibilities:
- Implemented reporting Data Warehouse with online transaction system data.
- Developed and maintained data warehouse for PSN project.
- Provided reports and publications to Third Parties for Royalty payments.
- Managed user account, groups and workspace creation for different users in Powercenter.
- Wrote complex UNIX/windows scripts for file transfers, emailing tasks from FTP/SFTP.
- Worked with PL/SQL procedures and used them in Stored Procedure Transformations.
- Extensively worked on oracle and SQL server. Wrote complex sql queries to query ERP system for data analysis purpose
- Worked on most critical Finance projects and had been the go to person for any data related issues for team members.
- Migrated ETL code from Talend to Informatica. Involved in development, testing and post production for the entire migration project.
- Documented the code.
- Tuned ETL jobs in the new environment after fully understanding the existing code.
- Maintained Talend admin console and provided quick assistance on production jobs.
- Involve in designing Business Objects universes and creating reports.
- Built adhoc reports using stand-alone tables.
- Involved in creating and modifying new and existing Web Intelligence reports.
- Created Publications which split into various reports based on specific vendor.
- Wrote Custom SQL for some complex reports.
- Worked with business partners internal and external during requirement gathering.
- Worked closely with Business Analyst and report developers in writing the source to target specifications for Data warehouse tables based on the business requirement needs.
- Exported data into excel for business meetings which made the discussions easier while looking at the data.
- Performed analysis after requirements gathering and walked team through major impacts.
- Provided and debugged crucial reports for finance teams during month end period.
- Addressed issue reported by Business Users in standard reports by identifying the root cause.
- Get the reporting issues resolved by identifying whether it is report related issue or source related issue.
- Creating Ad hoc reports as per users needs.
- Investigating and Analysing any discrepancy found in data and then resolving it.
Technologies Used: Informatica Power Center 9.1/9.0, Talend 4.x & Integration suite, Business Objects XI, Oracle 10g/11g, Oracle ERP, EDI, SQL Server 2005, UNIX, Windows Scripting, JIRA
Confidential, San Diego, CA
ETL Developer
Responsibilities:
- Gathered business requirements and prepared technical design documents, target to source mapping document, mapping specification document.
- Extensively worked on Informatica Powercenter.
- Parsed complex files through Informatica Data Transformations and loaded it to Database.
- Optimized query performance by oracle hints, forcing indexes, working with constraint based loading and few other approaches.
- Extensively worked on UNIX Shell Scripting for splitting group of files to various small files and file transfer automation.
- Worked with Autosys scheduler for scheduling different processes.
- Performed basic and unit testing.
- Assisted in UAT Testing and provided necessary reports to the business users.
Technologies Used: Informatica Power Center 8.6, Oracle 10g/11g, UNIX Shell Scripting, Autosys
Confidential, Irving, TX
ETL/Data Warehouse Developer
Responsibilities:
- Gathered requirements from Business and documented for project development.
- Coordinated design reviews, ETL code reviews with teammates.
- Developed mappings using Informatica to load data from sources such as Relational tables, Sequential files into the target system.
- Extensively worked with Informatica transformations.
- Created datamaps in Informatica to extract data from Sequential files.
- Extensively worked on UNIX Shell Scripting for file transfer and error logging.
- Scheduled processes in ESP Job Scheduler.
- Performed Unit, Integration and System testing of various jobs.
Technologies Used: Informatica Power Center 8.6, Oracle 10g, SQL Server 2005, UNIX Shell Scripting, ESP job scheduler
Confidential, St. Louis, MO
Data Analyst
Responsibilities:
- Involved in understanding the legacy applications & data relationships.
- Attended user design sessions, studied user requirements, completed detail design analysis and wrote design specs.
- Interacted with key users and assisted them with various data issues, understood data needs and assisted them with Data analysis.
- Prepared and maintained documentation for on-going projects.
- Worked with Informatica PowerCenter for data processing and loading files.
Technologies Used: Informatica PowerCenter 7.x, Oracle 9i, SQL, Toad, SQL Plus