Sr. Data Engineer Resume
San Jose, CA
SUMMARY
- Passionate professional with 12 years of experience, specializing in Big Data, AWS Stack, Data Warehousing Architecture, DataModeling, Analysis, Design, Development, Problem Solving, Performance Tuning and Technical Support.
- Ability to explore, experiment new technologies and quick adaptability on fast pace environment.
- Design and build Business Data Warehouse and other aspects of the Big Data Initiatives.
- Design and implement Hadoop framework and development in HDFS, MapReduce, Hive, Pig
- Good experience in Python and Spark.
- Strong experience in AWS major services like Redshift, EMR, EC2, S3, Lambda, DynamoDB and other
- Experience in design and orchestrating the data pipelines for data integration and ingestion using Airflow.
- Rich ETL experience in building reusable ETL framework and complex mappings using Informatica Cloud, Spark, Informatica PowerCenter or any ETL/ELT tools and institutionalizing ETL best practices, Methodologies, Re - usable components, Conventions and Standards.
- Expertise in troubleshooting and issue resolution related to system performance, ETL applications and performance tuning, BI application and data integrity
- Proficient in writing SQL, PL/SQL, Teradata Utils and query performance tuning.
- Ability to understand Business requirement and translate into technical/mapping specifications
- Highly motivated, strong analytical, planning, organizational and problem solving skills
- Excellent analytical, written & verbal communication, interpersonal skills and ability to effectively communicate conceptual and technical subjects.
TECHNICAL SKILLS
ETL: Informatica Cloud, Spark, Informatica PowerCenter 6.X/7.X/8.X/9.x, ODI 11.X, Decision Point Warehouse Application (DPA) 10.5.1, SSIS
AWS: Redshift, Lambda, EMR, EC2, S3, DynamoDB, RDS, Kinesis, ElastiCache, ElasticSearch, CloudFormation, CloudWatch, IAM, DataPipeline, and SNS etc…
Big Data: Hive, Pig, MapReduce, HDFS, Hadoop, HBase, NoSQL, MongoDB, Sqoop, Yarn
OLAP: OBIEEE 10.1.3.3.3 , Essbase 9.3.5,Hyperion Performance Suite (BRIO) 8.3.2, Business Objects 5.1/6.1/ XI R2, SSRS, SSAS
Database: Amazon Redshift, Snowflake, Teradata 7.X/12.X, Oracle 8i/9i/10g/11g, DB2, Sybase, SQL Server 2000/05/08, MySQL, MS Access
Language: C, C++, SQL, PL/SQL, T-SQL, Python, Perl, R
OS: OSX, Windows 7/98/2000/XP, UNIX/Linux
Others: Airflow, Jenkins, Presto, PyCharm, Sublime, Erwin 4.0, Tivoli, Control M, MS office, AccuRev, MS Visual Studio 2010, Oracle EBS Suite, BugZilla, HP QC, Service Now, DAC etc…
PROFESSIONAL EXPERIENCE
Confidential, San Jose, CA
Sr. Data Engineer
Responsibilities:
- Evaluate and set up the environments for various new tech stacks.
- Design and Implement AWS environment end-to-end solution for Cloud Migration.
- Worked and developed solutions in major AWS services like Lambda, EMR, EC2, Kinesis, Elastic Search, CloudFormation.
- Build new data pipelines for data integration and data ingestion using Spark/Python/Scala and Airflow.
- Developed ETL’s in using Spark SQL, RDD, Data Frames.
- Design and develop reusable components, frameworks and UDF’s for ingestion and data quality.
- Provide architectural solutions when building new and migrating existing applications, software and services on the AWS platform.
- Migrate the on-premise databases to Redshift database, DynamoDB, RDS - Mysql
- Airflow admin, design, develop and orchestrate the data pipelines for batch data processing
- CloudFormation, develop automation scripts for deployment, continuous integration and testing
- Informatica Cloud admin, design and develop the ETLs
- Python Developer, create the reusable classes, create reusable and dynamic Airflow Dags.
Environment: Informatica Cloud, Python, Airflow, Spark, Github, SQL Server, Hadoop, Hive, and Presto.
Confidential, San Jose, CA
Sr. Data Engineer
Responsibilities:
- Acquired knowledge of our multiple types of operations data with the desire to understand how the information is used by the business.
- Work in conjunction with Business Analyst, DBA, Data Architects on the backend data warehouse reporting solutions.
- Assist DBA in tuning SQL queries and stored procedures to improve performance.
- Manage and perform data cleansing, de-duplication and harmonization of data received from, and potentially used by, multiple systems.
- Engage in maintaining and troubleshooting daily data loads and addressing any issues.
- Work in OracleBI Applications, DAC and ETL repositories and execution plans
Environment: Informatica 9.5, Siebel, OBIEE 11.x, DAC, Oracle 11g, PL/SQL, SQL Server, SSIS, JIRA, Hadoop
Confidential, San Jose, CA
Data Engineer
Responsibilities:
- Translate business requirements into technical requirements.
- Responsible for entire BI database, and make sure data availability for end users on time.
- Designed ETL architecture and architecture documents and mapping documents.
- Established best practice and standard for ETL process documents.
- Investigate processes to determine the optimal method to implement changes while minimizing impact to existing systems.
- Interact with functional experts, at all levels, to understand business issues, challenges and identify new opportunities
- Instrumenting data pipelines on big data platform Hive, and creating UDF
- Work in implementing Hadoop environment and work on HDFS, Map Reduce, HBase
- Writing code on Pig Latin, HiveSQL for the data integration.
- Provide various ETL solutions by creating various components in Oracle Data Integrator (ODI) and also by writing PL/SQL, and Shell scripts.
- Creation of Reports, Dashboards using OBIEE and Hyperion
Environment: Oracle Data Integrator 11.x, Informatica 9.5, Bigdata- Hadoop, HDFS, MapReduce, Pig, Hive, HBase, OBIEE 11.x, Oracle 11g, Perforce P4V, BugZilla, Mysql
Confidential, San Diego, CA
Lead Developer
Responsibilities:
- Analyze, code, and test applications in both Informatica and SQL platforms. Lead the development and maintenance of products; participate in systems architecture maintenance, planning and oversight.
- Extensively worked on Informatica 9.0/8.6 to migrate data from various input sources like SQL Server, oracle, DB2 and flat files into SQL Server 2008 target.
- Respond to and assist with troubleshooting production application system and/or make necessary corrections and/or modification for resolution of issues.
- Document system procedures and guidelines for department and/or users. Produce technical documentation in accordance with SDLC guidelines.
- Have an active role in the development group by sharing knowledge and mentoring others.
Environment: Informatica 9.0/8.6, SQL Server 2008 R2, Oracle 10g, Quality Center 10, HP Service Manager, Visual Studio 2010, SQL Server Management Studio etc.
Confidential, Seattle, WA
Technical Lead/ETL Architect
Responsibilities:
- Work with business to gather requirements and define the solution.
- Design Dimension Modeling and Define Mapping specification rules.
- Analysis and Design enhancement in Informatica, Teradata utilities like B-Tech, M-Load, Fast Load, Parallel Transporter, T-Pump.
- Creation of ETL mappings and Reports & Dashboards.
Environment: Informatica 8.6.1, Teradata V2R6, AccuRev, Oracle, UNIX, Perl
Confidential, Milpitas, CA
Technical Lead/ETL Architect
Responsibilities:
- Business requirement analysis, Impact analysis.
- Lead programmer on implementation of packaged data warehouse solution for customer relationship management and supply chain management modules.
- Responsible for estimation of effort and preparation of Task Plan for the offshore team
- Analysis and Design enhancement in Informatica, DPA, ODI, BRIO, OBIEE, Essbase, Oracle EBS suite, and UNIX for requirements.
- Building RPD with appropriate joins and tables in Physical, BMM layers and creation of Reports, Dashboards using OBIEE and BRIO
- Responsible for data integrity as well as the quality of data.
- User Acceptance Testing and implementation, change management activities and support
- Involved in production support duties to ensure all jobs run successfully to meet SLAs and also make sure users receive necessary reports on time.
Environment: Informatica PowerCenter 8.1.1, DPA 10.5.1, BRIO 8.3.2, SAP BO, OBIEE 10.1.3.3.3 , DAC, Oracle EBS Suite, Hyperion Essbase 9.3.5, Planning, Oracle, UNIX
Confidential
ETL Lead/ Ware House Analyst
Responsibilities:
- Gathering the requirement and involved in total life cycle of the project
- Review/Prepare the Source maps to fit according to FS-LDM
- Develop numerous Informatica mappings to implement ETL logic
- Wrote B-teq scripts for complex extraction logic mappings
- Lead the PIR team with the size of 5 and delivery manager for the product.
Environment: Terada V2R5, Informatica 7.1.2, UNIX
Confidential
ETL Lead/ Ware House Analyst
Responsibilities:
- Work with business to gather requirements and define the solution.
- Estimation of effort and preparation of Task Plan, preparation of SOW
- Analysis and Design enhancement in DB2 scripts, Informatica mappings, UNIX, Data mart application fine tuning
- Creation of Informatica mappings, Teradata scripts
- User Acceptance Testing and implementation, change management activities and support
- Involved in production support duties to ensure all jobs run successfullyto meet SLAs and also make sure users receive necessary reports on time
Environment: DB2, Informatica 7.1.2, UNIX
Confidential
ETL Lead/ Ware House Analyst
Responsibilities:
- Gathering and getting clarified the requirements from the business users
- Setup project environment in Development & Testing
- Wrote numerous B-teq scripts for complex extraction logic mappings.
- Programing the project and plan the task and leading the team
- Responsible to get project UAT sign off, deployment and initial production support
Environment: Teradata V2R5, Informatica 7.1.2, Business Objects XI, Windows XP, UNIX.
Confidential
ETL Developer
Responsibilities:
- Developing ETL mappings in Informatica for Full and Incremental loads.
- Informatica Administration includes taking repository backup.
- Configuring of DAC Client for Scheduling.
- Design BI reporting solutions. Hands-on on Web Intelligence, Desktop Intelligence, Designer.
- Creating the Logical and Physical joins, hierarchies in Siebel Analytics Business Layer.
- Creating the reports with embedding the Global Filters/Column selectors, charts.
Environment: Oracle 9i, Informatica 7.1.2, DAC, Siebel Analytics 7.8.4, Windows XP, UNIX.