We provide IT Staff Augmentation Services!

Etl Designer/hadoop Developer Resume

5.00/5 (Submit Your Rating)

El Segundo, CA

PROFESSIONAL SUMMARY

  • Over 10 plus years of IT experience with 8 1/2 years experience in data migration, data warehousing database design, Analysis and Maintenance projects.
  • Over 1 year of experience implementing big data analytic solutions using HADOOP HDFS, MAP REDUCE, HIVE, PIG, SQOOP, and OOZIE.
  • Extensiveexperience in database design for data warehouses with strong understanding of Dimensional Modeling, Star and SnowflakeSchemas combined with data staging, database design, datawarehouse requirement analysis and Implementation.
  • Experience in planning, architectural design and optimization ofTeradatadata warehousing.
  • Hands on Experience in ETL, OLAP, Data Analysis, Design, Development, Implementation, Testing and trouble shooting of Business Intelligence, Data Warehousing, and Database applications.
  • Very good understanding of Teradata MPP architecture such as Shared Nothing, Nodes, AMPs, BYNET, Partitioning, Primary Indexes etc. Extensively used different features of Teradata such as Fast Load, Multi Load, BTEQ scripting, Fast Export and SQL Assistant.
  • Exclusive experience in providing Business intelligence solutions using Informatica products such as Power Center 9.5/9.1/8.6.1/8.5.1/8.1.1/8.0/7.1, Power mart 6.1, Power Exchange for data models and OBIEE 10.1.x.
  • Strong experience in performing ETL operations like Data Extraction, Data Transformation and Data Loading with Informatica Power Center and Informatica Power Mart (Repository Manager, Designer, Server Manager, Workflow Manager, and Workflow Monitor).
  • Experience in designing and developing complex mappings from varied transformation logic like Unconnected and Connected lookups, Source Qualifier, Router, Filter, Expression, Aggregator, Joiner, Update Strategy etc.
  • Extensively worked on data extraction, Transformation and loading data from various sources like Oracle, SQL Server and Flat files.
  • Hands on experience in tuning mappings, identifying and resolving performance bottlenecks in various levels like sources, targets, mappings and sessions.
  • Extensive experience in Business Analysis, Data Analysis & clarification of requirements with SME (SubjectMatter Expert). Created documentation of Business process and implemented the same.
  • Extensively worked on designing Business views & writing SQL queries in Teradata catering specific business requirement to generate data.
  • Experienced in performance tuning of user queries by analyzing the explain plans, recreating the user driver tables by right primary Index, scheduled collection of statistics, secondary or various join indexes.
  • Developed UNIX shell scripts and used BTEQ, Fast Load, Multi Load, and Fast Export utilities extensively to load to target database.
  • Used extensively Derived Tables, Volatile Table and GTT tables in many of the BTEQ scripts.
  • Experience in integration of various data sources with Multiple Relational Databases like Teradata, Oracle, and SQLServer.
  • Proficient using in RDBMS like Oracle, SQL Server, Teradata, PL/SQL, SQL *PLUS, SQL *Loader and TOAD.
  • ExperienceinInformatica(versions9.5), Mappings,Transformations,WorkflowManager,

TECHNICAL SKILLS

Database: Teradata 14/13, V2R5/V2R6.0, Oracle (8i/9i/10g), SQL Server 2000

Teradata tools & Utilities query facilities: SQL Assistant, BTEQ

Load & Export: Fast Load, Multi Load, Fast Export, BTEQ, Oracle SQL Loader

ETL: Informatica (8,9)

Big Data: Hadoop, Map reduce, Hive, Pig, Oozie, and Sqoop

Languages: C, C++, Visual Basic

Scripting languages: UNIX Shell, Perl,Python

Replication Tool: Golden Gate, Data Mover

Others: Business Objects 6.5,OBIEE, Kintana, Quality Center, Defect Tracker

PROFESSIONAL EXPERIENCE

Confidential, El Segundo, CA

ETL Designer/Hadoop Developer

Responsibilities:

  • Extract, Transform and Load data from various relational data sources like Teradata, Oracle, and SQL server into Teradata data warehouse using Informatica power center applications.
  • Created mapplets/re - usable transformations for further use in mappings.
  • Created generic UNIX scripts to handle daily file check routines, FTP transfers, and data
  • cleansing operations.
  • Worked on most of the commonly used mapping designer transformations like lookup, joiner,
  • aggregator, router, normalizer, update strategy, expression, source qualifier, sorter, sequence generator.
  • Created complex ETL workflows and scheduled jobs in Autosys.
  • Involved in unit testing/iterative testing/preparation of test cases for ETL load jobs.
  • Implemented Teradata Fastload control scripts to load flat files data into staging area.
  • Implemented FastExport control scripts to export data from Teradata tables.
  • Implemented Multiload control scripts to handle updates/inserts/upserts into warehouse tables.
  • Created complex SQL scripts using Teradata BTEQ.
  • Performance tuning of informatica load jobs/Teradata BTEQ batch scripts.
  • Created Teradata Parallel Transporter scripts to copy the data from one environment to thether environments.
  • Implemented Informatica & Teradata best practices at the work site.
  • Acquisition of structured data into HDFS from Teradata using SQOOP utility.
  • Exported HDFS data to relational databases using SQOOP for visualization and to generatereports for BI team.
  • Created HIVE internal/external tables for handling daily Hadoop ETL processes.
  • Implemented partitioning & bucketing concepts while creating HIVE tables.
  • Created OOZIE workflows to handle the scheduling of multiple HIVE, PIG, and SQOOPscripts.
  • Adhoc analysis of HDFS/HIVE data using Impala.
  • Performance tuning of HIVE queries.
  • Implemented business logic and transformations using HIVE UDF’s and regular expressions.
  • Monitoring of HADOOP map/reduce jobs using cloudera manager and troubleshoot long running processes.
  • Work with the business and IT in respect to the data interpretation and control of current reporting and information delivery.
  • Follow agile methodology procedures to implement better ETL solutions and thereby reduce post-production bugs.
  • Gather business requirements and perform data analysis.
  • Identify production discrepancies and provide solutions to resolve them.
  • Design logical/physical/dimensional data models to define new data structures that are later materialized into physical tables on the database.
  • Design architecture level/table level/column level specs for development phase.
  • Developed mapping documents for loading the data from source to target.

Confidential, San Jose, CA

Senior Teradata Lead Developer/Technical Analyst

Responsibilities:

  • rWorked with business users to identify, understand and build the data model.
  • Interacted with business analysts, data architects, application developers to develop a data model.
  • Lead the ETL team technically.
  • Responsible for providing the timely status reports, addressing technical issues andfacilitating design.
  • Performance tuning of the reports queries.
  • Developed UNIX scripts to load materialized tables in ETL.
  • Recommend and Implement best practices in Teradata/Informatica
  • Track performance of report on daily basis.

Confidential, San Francisco, CA

Senior Teradata Lead Developer/Technical Analyst

Responsibilities:

  • Recommend and Implement best practices in Teradata.
  • Writing Teradata SQL queries to join or any modifications in the table.
  • Tuned the existing SQL and achieved increased performance for faster user query performance.
  • Resolving performance bottlenecks in various levels like sources, targets, mappings and sessions.
  • Created and Modified Indexes, PPI, Statistics, Character Sets, Materialized tables etc.
  • Created proper PI taking into consideration both planned access and even distribution of data across all the available AMPS.
  • Acted as a single resource with sole responsibility of Teradata.
  • Requirement gathering of VMI report (Sears and Kohl’s) from Users.
  • Worked with ETL and Business Object resources to implement the end-to-end change.
  • Involved in Unit Testing and Preparing test cases and also involved in Peer Reviews.
  • Used Teradata Viewpoint to monitor the warehouse's Teradata infrastructure and report out tomanagement root cause analysis or suggest workarounds for production incidents that occur.
  • Used in-depth expertise in the Teradata cost based query optimizer, identified potential bottlenecks with queries from the aspects of query writing, skewed redistributions, join order, optimizer statistics, physical design considerations (PI and USI and NUSI and JI etc) etc.
  • Involved in production deployment and post deployment support.
  • Developed UNIX scripts to load materialized tables in ETL.
  • Track performance of report on daily basis.
  • Keep track of the VMI report for Sears and Kohl’s.

Confidential, San Jose, CA

Senior Teradata Lead Developer-ETL

Responsibilities:

  • Designed, developed, tested, enhanced and maintained Teradata based Enterprise DataWarehouse, using Teradata, Teradata SQL, Teradata ETL utilities, Informatica, UNIX and Teradata Analytical functions.
  • Designed, developed, tested, optimize and maintained processes that involve acquisition/extraction, transformation and loading (ETL/ELT) of Tera-byte volumes of data.
  • Data extraction, Transformation and loading data from various sources like Oracle and Flat files.
  • Worked with Datawarehouse Architects on LDM/PDM for ongoing projects.
  • Coordinated with business owners to gather requirements and analyzingdata regarding current functionality and operational adequacy of existing EDW to determine recommend and develop various modifications.
  • Worked with DBA team to ensure implementation of the databases for the physical data models intended for the data marts. Created proper Teradata Primary Indexes (PI) taking into consideration of both planned access of data and even distribution of data across all the available AMPS.
  • Helped Business Users by writing Complex efficient Teradata SQLs to get a detailed for Data Mining.
  • Performance Tuning ETLs as well as reporting queries on Teradata EDW in Cisco, which are run through various,subscribing business applications.
  • Developed UNIX shell scripts and used BTEQ, FastLoad, Multiload, and Fast Export utilities extensively to load to target database.
  • Employed good design methodologies to design, enhance, improve or otherwise develop systems throughout project life cycle in accordance with corporate and/or client functional needs, requirements, quality controlparameters and standard operating procedures
  • Data modeling for new subscriptions and applications as per need basis.
  • Worked on Audits implementation for meeting SOX requirements in the system.
  • Capacity Monitoring on the Production system with generation of Daily and Weekly reports.
  • Analysis and forecasting the performance activities.
  • Development and Tuning of ETL processes in Informatica and Teradata for various Data Marts.
  • Interacted with various Business teams for new subscriptions on Teradata EDW and supporting theCosting evaluation.
  • Worked on Informatica workflow scheduler and GUI scheduler.
  • Created UNIX Scripts to manipulate and Load the data.
  • Interacted with the DBA team and Teradata GSC team on various issues representing Cisco EDW.
  • Did Project Management activities as per the Client requirements by representing the team in various forums within the Organization as well as allocating tasks to the offshore lead and periodically assessing the task progressand conforming the same within the stipulated timelines.
  • Worked on the Global Delivery Model (Onsite - Offshore) for Production Support team.
  • Provided 24X7 support to the database activities in case of any errors or issues.

Confidential, San Jose, CA

Replication Solution Analyst/ Teradata Lead Developer

Responsibilities:

  • Designed, developed, tested and maintained disasterrecoverysystem forTeradata based Enterprise DataWarehouse using replication tools like Golden Gate and Data Mover.
  • Created Golden Gate parameter files to replicate the data to target database.
  • Performed data volume analysis to ensure 100% data replication.
  • Optimized SQL query to improve the performance of replication of Tera-byte volumes of data.
  • Provided replication solution defects resolution ideas/fix to Production support team.
  • Interacted with the Golden Gate platform team on various replication solution issues.
  • Worked on Audits implementation for disaster recovery system.
  • Worked with business users for gathering and developing latency dashboard for disaster recovery system.
  • Worked on the Global Delivery Model (Onsite - Offshore) for Architecture Team.
  • Monitor progress in all cross-functional areas according to milestones and schedules.
  • Interacted with technical and business analyst, operation analyst to resolve data issues.
  • Worked on complex queries to map the data as per the requirements.
  • Performed load-balancing analysis to move applicationto replication system.

Confidential, San Jose, CA

Teradata Developer

Responsibilities:

  • Developed and supported the Extraction, Transformation, and load process (ETL) for data migration using Informatica power center.
  • Data extraction, Transformation and loading data from various sources like Oracle and Flat files.
  • Understanding the business processes of Cisco and Requirement analysis.
  • Coordination with Business owners to gather Requirements
  • Facilitating Business Owners in Decision making by Data Analysis and Data Profiling
  • Helping offshore development teams in Requirements understanding and development
  • Assisting in Modeling to incorporate new requirements.
  • Created Informatica mappings/workflows and tuned the same for performance by implementation of Push Down
  • Optimization and strategic placement of collect stats.
  • Worked on Informatica workflow scheduler and GUI scheduler.
  • Designed and developed time-bound Informatica processes
  • Writing the Unix and BTEQ scripts for implementation of the History Fix.
  • Creation of BTEQ, Fast export, MultiLoad, TPump, Fast load scripts.
  • Used Data profiler to allow the analysis of data directly in the database, which improves performance, while eliminating the time and costs of moving data among databases.
  • Performance tuned Teradata SQL queries by using query plan and creating indexes to minimize query execution times.
  • Assisting the QA team in Testing and Dev to QA migration of History and Ongoing code. Also assisted the QAteam to get the Business and IT signoff for my stream.
  • Creation of BO Universe for Teradata environment.
  • Creating BO reports in Development environment and testing it in TEST environment.
  • Created, updated and maintained ETL technical documentation.
  • Working on the Global Delivery Model (Onsite - Offshore) for Architecture Team.
  • Assisting the Production Control team during PROD turnover and providing warranty support for any production issues that may arise
  • Assisting the Business Teams to generate reports and meet the requirements.

Confidential

Software Engineer

Responsibilities:

  • Informatica Power Center was used to extract data from various sources like flat files, excel files to load into the OracleData warehouse.
  • Created,updated and maintained ETL technical documentation.
  • Developing PL/SQL queries and objects using Oracle 8i
  • Developed test cases and performing a check on test cases list.
  • Development of Informatica mappings, sessions and workflows.
  • Created BO Universe for Oracle environment.
  • Created BO reports in Development environment and testing it in TEST environment.
  • Developed Unix script to load the data into Target database.
  • Coordinated with Onsite lead to gather Requirements.

We'd love your feedback!