Lead Data Warehouse Developer Resume
SUMMARY
- Highly skilled and dynamic professional with around 9 years of experience in Data warehousing domain.
- As a Data warehouse consultant/specialist, Strong noledge of teh Data Warehousing Techniques, Hands - on experience in design of mission critical applications in Enterprise Data Warehouse.
- Extensively worked on Teradata database platform like database design, data modeling, application design, development, enhancements, upgrades and support.
- Expert in writing SQL queries using Teradata and performance tuning and optimization of various applications. Secondary skill set includes Informatica, Hadoop and Mainframes.
- Excellent noledge on design and development of Data Mart /Data Warehousing applications using Teradata, Informatica, Hadoop and Unix Shell scripts.
- Strong noledge in Teradata architecture & utilities (BTEQ, FASTLOAD, FASTEXPORT, MULTILOAD, TPUMP, TPT & SQL Assistant, Teradata Manager, PMON etc.).
- Data modeling noledge using Dimensional Data Modeling, Star Schema Modeling, Snow Flake Modeling, FACT and Dimensions Tables.
- Extensively worked on Informatica Designer Components -- Source Analyzer, Warehouse Designer, Transformation Developer, Mapplet and Mapping Designer.
- Strong Experience on Workflow Manager Tools -- Task Developer, Workflow, Worklets Designer and Workflow Monitor. Used Both Pipeline Parallelism and Partition Parallelism for improving performance.
- Extensive experience in Extraction, Transformation and Loading of data from different heterogeneous source systems like Flat files, Oracle, SQL Server.
- Good skills in UNIX shell scripting.
- Hands-on experience with Productionizing Hadoop applications such as administration, configuration management, monitoring, debugging, and performance tuning.
- Complete understanding of Hadoop cluster setup. Installation, configuration and tuning on various sized configurations for development and testing.
- Experience in analyzing data using Hive QL and custom Map Reduce programs in Java. Has experience in ETL using DMExpress tool.
- Good experience in Designing, implementing, and improving analytic solutions for Big Data on Apache Hadoop, Hive and Sync sort DMExpress tool.
- Very good at preparing a neat HLD, LLD, Test plan, Test scripts, DTS, Run book and other related project documents.
- Extensive noledge in ETL Analysis, Design, Development, Testing, Implementation, Maintaining Standards, Quality Audits, Performance Tuning, Automation of jobs, building reusable components, Maintenance and support of various applications.
- Excellent skills in a wide variety of technologies and a proven Ability and willingness to learn new technology and deliver results in a timely manner.
- Very good communication skills and quick adaptability to new technologies and new working environment.
- Excellent organizational skills and ability to prioritize workload. Imparted various training sessions on Data warehousing Basics.
- A well-organized, goal-oriented, highly motivated TEMPeffective team member.
TECHNICAL SKILLS
Databases: Teradata V12/13.10, Oracle 10g/11g, SQL Server 2005/08/12 and DB2 6/8
Data warehousing ETL tools: Informatica 9.x/8.x/7.x and Hadoop 1.0/2.0
Operating System: Windows, Z/OS (Mainframe) and UNIX.
Scheduling tools: Autosys, CA7, Control-M and DMX
Languages: UNIX Shell Scripting and C, C++, JAVA, HTML and CSS
Business Areas: Banking and Financial services - Deposits, Investments, Small business and commercial loans and GL Segm commercial loans and GL Segmentation
Area Specialization: Informatica, Teradata, Hadoop Hive, DMExpress, UNIX, JCL & DB2
Modeling Tools: Erwin
PROFESSIONAL EXPERIENCE
Confidential
Lead Data warehouse Developer
Environment: UNIX, Sync Sort DMExpress Tool, TeradataV13.10, Informatica Power Center 9.5/9.1, Hadoop and Teradata SQL Assistant
Responsibilities:
- Participate in project scope identification, data architecture development, and project planning.
- Assist with workload prioritization and management.
- Interact with Users and Business Analysts on a daily basis for requirements gathering, clarifications and suggestions in application design.
- Analyze existing system flow in Informatica using designer components to understand end to end data flow and draw proposed data flow in new ETL platform (DMExpress - Hadoop).
- Establish procedures to ensure integrity and responsiveness. Prepare HLD, LLD, Test plan and test cases. Design walk-through with all stake holders and finalize.
- Develop a code in DMExpress (GUI tool to connect Hadoop cluster - HDFS, teh cheap commodity hardware) using Map Reduce to convert student loan ETL process from Informatica platform to DMExpress-Hadoop. Communicate data issues to other groups/vendors in a timely manner.
- Developed complex Map Reduce jobs in DMExpress tool such as change data capture (capture new, changed, unchanged and purged records), huge transformation logics etc using DMX job and Task tools.
- Migrated costly & complex ETL logics from Teradata such as high CPU OLAP queries to DM Express Map reduce jobs to save CPU and space on expensive Teradata platform.
- Engage source team to push teh operational data to new platform.
- Develop data lineage for new platform. Build teh test environment and move teh legacy data across platforms using Teradata utilities like TPT, fast load, and fast export for testing & data validation and ensure no change in data after migrating to new platform.
- Assure all programs are documented using standardized documentation procedures.
- Executed Direct data flow, Map side join and Reducer side join per requirement for better CPU utilization.
- Code review with all stake holders to ensure all requirements covered and code is adhered to enterprise standards and guidelines.
- Parameterized teh layout tasks, dmx jobs, tasks for readability and reusability.
- Scheduling teh jobs in Autosys and Implement teh code in Production. Observe teh application behavior in production till warranty and resolve if any issues.
- Establish archival and disaster recovery procedures to ensure data preservation.
Confidential
Sr Data warehouse Developer/Analyst
Environment: UNIX, TeradataV13.10, Oracle 11g, Informatica Power Center 9.1, Hadoop, Java Map Reduce, Hive QL and Teradata SQL Assistant
Responsibilities:
- Participated in Project scope and Estimation and planning. Involved in Source data analysis, Database design, Modeling, Metadata build and data lineage build.
- Data modeling includes Normalized model (for data consistency) and de-normalized model (for speedy response to users).
- Daily interaction with Users and Business Analysts for requirements gathering, clarifications and suggestions in application design.
- Engage all teh stake holders for necessary design confirmations and to understand business needs.
- Prepare source to target mapping document (data lineage) based on design assumptions.
- Understand BRD and prepare HLD, LLD, Test plan and test cases. Coordinating with respective stake holders to resolve SIT/UAT Defects.
- Extract, cleanse teh data(which usually be received to UNIX platform from operational systems) and transform data using Hive QL, Oozie XML based on agreed data lineage by all stake holders and ensure it is accurate, reliable, consistent and available for user analytics.
- Implement Teradata tables (DDLS, Indexes, Grants space etc.) in lower environments per design and verify if proposed model works well in real time.
- Develop a ETL code in Hadoop using Java Map reduce & Hive QL to parse and process practice loan file and load teh data to Teradata environment for analytics by vast user community.
- Experience in managing and reviewing Hadoop log files (aka job tracker and task tracker).
- Importing and exporting data into HDFS and Hive/Teradata using SQOOP.
- Parameterized teh shell and Oozie XML scripts and increased teh readability & re-usability.
- Used Teradata BTEQ, MLOAD utilities to move staging data to enterprise data warehouse.
- Scheduling teh jobs in Autosys and Implement teh code in Production. Observe teh application behavior in production till warranty and resolve if any issues.
- Establish archival and disaster recovery procedures to ensure data preservation.
Confidential
Sr Data warehouse Developer
Environment: UNIX, Oracle 11g, TeradataV13.10, Teradata SQL Assistant and Informatica Power Center 9.1
Responsibilities:
- Created data lineage document (source to Target mapping) based on teh load control rules framed by business users.
- Prepare HLD, LLD with data flow diagrams and business requirements. Design walk through with teh internal and external stake holders of teh project.
- Used Informatica Power Center 9.1 for extraction, transformation and load (ETL) of data in teh data warehouse.
- Load and transform large sets of structured, semi structured and unstructured data
- Extensively used Transformations like Source Qualifier, Filter, Router, Aggregator, Sorter, Joiner, Expression and Lookup, Update strategy and Sequence generator and SQL Transformations to perform load control checks.
- Developed complex mappings in Informatica to apply all load control checks for TEMPeffective pre and post processing of ETL and build centralized data model to store control pass/fail results for all retail loan applications.
- Implemented performance tuning logic on targets, sources, mappings, sessions to provide maximum efficiency and performance.
- Parameterized teh mappings and increased teh re-usability.
- Used Informatica Power Center Workflow manager to create sessions, workflows and batches to run with teh logic embedded in teh mappings.
- Provided inputs for teh most optimum and feasible design in terms of performance and maintenance after thorough analysis of teh requirement
- Scheduled teh jobs in Autosys and did monitoring its daily run in Production. Maintain teh applications till warranty.
- Analyzed production support related issues, inquiries and requests and provided a concrete explanation of resolutions to issues, inquiries and requests.
Confidential
Data warehouse Developer
Environment: Oracle 11g, Informatica 8.6.1, UNIX, Autosys, TeradataV12, Teradata SQL Assistant and Mainframe JCL
Responsibilities:
- Daily interaction with clients and Business Analysts for requirements gathering, clarifications and suggestions in data modeling and application design.
- Created data lineage document (source to Target mapping) based on teh requirements framed by business users.
- Used Teradata for data extraction, transformation and load (ETL) of data in teh data warehouse.
- Extensively used Teradata SQL and its OLAP functions such as RANK, ROW NUMBER, CSUM and MSUM etc to perform ETL.
- Developed complex mappings to capture customer data from various sources which lies within enterprise data warehouse to capture customers based on certain parameters and segregate them based on teh needs of business analytic users.
- Implemented performance tuning everywhere to ensure ETL runs smooth with increased customer volume day by day.
- Parameterized teh components and increased teh readability and re-usability.
- Used Mainframe JCL to create jobs, control cards to run Teradata SQL.
- Code Walkthrough and Implement corrections and suggestions before migrating to higher environments.
- Build control plan document for every enhancement for better project tracking and to ensure error free installs and application turnover. Provided inputs for teh most optimum and feasible design in terms of performance and maintenance after thorough analysis of teh requirement.
- Scheduled teh jobs in CA7 and did monitoring its daily run in Production. Maintain teh applications till warranty.
- Analyzed production support related issues, inquiries and requests and Provided a concrete explanation of resolutions to issues, inquiries and requests
Confidential
Application Developer
Environment: TeradataV12, Informatica 8.1, Oracle 10g, Teradata SQL Assistant, UNIX and Mainframe JCL.
Responsibilities:
- Mapping Data Items from Source Systems to teh Target System.
- Preparing HLD, LLD based on Business requirements preparing Test plan, Unit Test Cases per each requirement in BRD and coordinating with respective source teams for fixing SIT and UAT Defects.
- Used Teradata to create mappings using transformations to move data to a data Warehouse.
- Involved in requirements gathering, creating technical specifications document, developing mappings, testing teh mappings, providing support.
- Extensively used Teradata Fast Load, MultiLoad and T-Pump to transfer large volumes of data to BDW model.
- Interact with all teh stakeholders for requirements gathering, clarifications and suggestions in data lineage and application design.
- Predominantly focused on building a warehouse model dat can fit into bank standards and requirements. Remediate all teh queries to ensure performance is improved. Avoid usual problems in data warehousing environment like redundant data across teh systems in warehouse, Inconsistent data (not normalized) and occupy huge space etc.
- Scheduling teh jobs in CA7 & Autosys and monitoring its daily run in Production.
- Implementing teh components and maintain teh application till warranty.
Confidential
Application Developer
Environment: TeradataV2R5, Informatica 7.1/8.1, Oracle 9i/10g, Teradata SQL Assistant, UNIX and Mainframe JCL.
Responsibilities:
- Understanding Confidential India reward requirements (HLD/LLD) and preparing all teh necessary documents like run book, Low Level Design and DTS.
- Developed reward programs in mainframe Teradata and JCL for Confidential India credit cards.
- Preparing Unit Test Cases and Unit Test Results. Coordinating with respective source teams for fixing SIT and UAT Defects.
- Scheduling teh jobs in CA7 and monitoring its daily run in Production.
- Implementing components through changeman package and resolve if any issues post deployment.
- Mapping Data Items from Source Systems to teh Target System.
- Preparing HLD, LLD based on Business requirements preparing Test plan, Unit Test Cases per each requirement in BRD and coordinating with respective source teams for fixing SIT and UAT Defects.
- Used Teradata to create mappings using transformations to move data to a data Warehouse.
- Involved in requirements gathering, creating technical specifications document, developing mappings, testing teh mappings, providing support.
- Extensively used Teradata Fast Load, MultiLoad and T-Pump to transfer large volumes of data to BDW model.
- Interact with all teh stakeholders for requirements gathering, clarifications and suggestions in data lineage and application design.
- Parameterized teh components and increased teh readability and re-usability.
- Used Mainframe JCL to create jobs, control cards to run Teradata SQL.
- Code Walkthrough and Implement corrections and suggestions before migrating to higher environments.
- Build control plan document for every enhancement for better project tracking and to ensure error free installs and application turnover. Provided inputs for teh most optimum and feasible design in terms of performance and maintenance after thorough analysis of teh requirement.