Lead Data Warehouse Developer Resume
SUMMARY
- Highly skilled and dynamic professional with around 9 years of experience in Data warehousing domain.
- As a Data warehouse consultant/specialist, Strong knowledge of the Data Warehousing Techniques, Hands - on experience in design of mission critical applications in Enterprise Data Warehouse.
- Extensively worked on Teradata database platform like database design, data modeling, application design, development, enhancements, upgrades and support.
- Expert in writing SQL queries using Teradata and performance tuning and optimization of various applications. Secondary skill set includes Informatica, Hadoop and Mainframes.
- Excellent knowledge on design and development of Data Mart /Data Warehousing applications using Teradata, Informatica, Hadoop and Unix Shell scripts.
- Strong knowledge in Teradata architecture & utilities (BTEQ, FASTLOAD, FASTEXPORT, MULTILOAD, TPUMP, TPT & SQL Assistant, Teradata Manager, PMON etc.).
- Data modeling knowledge using Dimensional Data Modeling, Star Schema Modeling, Snow Flake Modeling, FACT and Dimensions Tables.
- Extensively worked on Informatica Designer Components -- Source Analyzer, Warehouse Designer, Transformation Developer, Mapplet and Mapping Designer.
- Strong Experience on Workflow Manager Tools -- Task Developer, Workflow, Worklets Designer and Workflow Monitor. Used Both Pipeline Parallelism and Partition Parallelism for improving performance.
- Extensive experience in Extraction, Transformation and Loading of data from different heterogeneous source systems like Flat files, Oracle, SQL Server.
- Good skills in UNIX shell scripting.
- Hands-on experience with Productionizing Hadoop applications such as administration, configuration management, monitoring, debugging, and performance tuning.
- Complete understanding of Hadoop cluster setup. Installation, configuration and tuning on various sized configurations for development and testing.
- Experience in analyzing data using Hive QL and custom Map Reduce programs in Java. Have experience in ETL using DMExpress tool.
- Good experience in Designing, implementing, and improving analytic solutions for Big Data on Apache Hadoop, Hive and Sync sort DMExpress tool.
- Very good at preparing a neat HLD, LLD, Test plan, Test scripts, DTS, Run book and other related project documents.
- Extensive knowledge in ETL Analysis, Design, Development, Testing, Implementation, Maintaining Standards, Quality Audits, Performance Tuning, Automation of jobs, building reusable components, Maintenance and support of various applications.
- Excellent skills in a wide variety of technologies and a proven Ability and willingness to learn new technology and deliver results in a timely manner.
- Very good communication skills and quick adaptability to new technologies and new working environment.
- Excellent organizational skills and ability to prioritize workload. Imparted various training sessions on Data warehousing Basics.
- A well-organized, goal-oriented, highly motivated effective team member.
TECHNICAL SKILLS
Databases: Teradata V12/13.10, Oracle 10g/11g, SQL Server 2005/08/12 and DB2 6/8
Data warehousing ETL tools: Informatica 9.x/8.x/7.x and Hadoop 1.0/2.0
Operating System: Windows, Z/OS (Mainframe) and UNIX.
Scheduling tools: Autosys, CA7, Control-M and DMX
Languages: UNIX Shell Scripting and C, C++, JAVA, HTML and CSS
Business Areas: Banking and Financial services - Deposits, Investments, Small business and commercial loans and GL Segm commercial loans and GL Segmentation
Area Specialization: Informatica, Teradata, Hadoop Hive, DMExpress, UNIX, JCL & DB2
Modeling Tools: Erwin
PROFESSIONAL EXPERIENCE
Confidential
Lead Data warehouse Developer
Responsibilities:
- Participate in project scope identification, data architecture development, and project planning.
- Assist with workload prioritization and management.
- Interact with Users and Business Analysts on a daily basis for requirements gathering, clarifications and suggestions in application design.
- Analyze existing system flow in Informatica using designer components to understand end to end data flow and draw proposed data flow in new ETL platform (DMExpress - Hadoop).
- Establish procedures to ensure integrity and responsiveness. Prepare HLD, LLD, Test plan and test cases. Design walk-through with all stake holders and finalize.
- Develop a code in DMExpress (GUI tool to connect Hadoop cluster - HDFS, the cheap commodity hardware) using Map Reduce to convert student loan ETL process from Informatica platform to DMExpress-Hadoop. Communicate data issues to other groups/vendors in a timely manner.
- Developed complex Map Reduce jobs in DMExpress tool such as change data capture (capture new, changed, unchanged and purged records), huge transformation logics etc using DMX job and Task tools.
- Migrated costly & complex ETL logics from Teradata such as high CPU OLAP queries to DM Express Map reduce jobs to save CPU and space on expensive Teradata platform.
- Engage source team to push the operational data to new platform.
- Develop data lineage for new platform. Build the test environment and move the legacy data across platforms using Teradata utilities like TPT, fast load, and fast export for testing & data validation and ensure no change in data after migrating to new platform.
- Assure all programs are documented using standardized documentation procedures.
- Executed Direct data flow, Map side join and Reducer side join per requirement for better CPU utilization.
- Code review with all stake holders to ensure all requirements covered and code is adhered to enterprise standards and guidelines.
- Parameterized the layout tasks, dmx jobs, tasks for readability and reusability.
- Scheduling the jobs in Autosys and Implement the code in Production. Observe the application behavior in production till warranty and resolve if any issues.
- Establish archival and disaster recovery procedures to ensure data preservation.
Confidential
Sr Data warehouse Developer/Analyst
Environment: UNIX, TeradataV13.10, Oracle 11g, Informatica Power Center 9.1, Hadoop, Java Map Reduce, Hive QL and Teradata SQL Assistant
Responsibilities:
- Participated in Project scope and Estimation and planning. Involved in Source data analysis, Database design, Modeling, Metadata build and data lineage build.
- Data modeling includes Normalized model (for data consistency) and de-normalized model (for speedy response to users).
- Daily interaction with Users and Business Analysts for requirements gathering, clarifications and suggestions in application design.
- Engage all the stake holders for necessary design confirmations and to understand business needs.
- Prepare source to target mapping document (data lineage) based on design assumptions.
- Understand BRD and prepare HLD, LLD, Test plan and test cases. Coordinating with respective stake holders to resolve SIT/UAT Defects.
- Extract, cleanse the data(which usually be received to UNIX platform from operational systems) and transform data using Hive QL, Oozie XML based on agreed data lineage by all stake holders and ensure it is accurate, reliable, consistent and available for user analytics.
- Implement Teradata tables (DDLS, Indexes, Grants space etc.) in lower environments per design and verify if proposed model works well in real time.
- Develop a ETL code in Hadoop using Java Map reduce & Hive QL to parse and process practice loan file and load the data to Teradata environment for analytics by vast user community.
- Experience in managing and reviewing Hadoop log files (aka job tracker and task tracker).
- Importing and exporting data into HDFS and Hive/Teradata using SQOOP.
- Parameterized the shell and Oozie XML scripts and increased the readability & re-usability.
- Used Teradata BTEQ, MLOAD utilities to move staging data to enterprise data warehouse.
- Scheduling the jobs in Autosys and Implement the code in Production. Observe the application behavior in production till warranty and resolve if any issues.
- Establish archival and disaster recovery procedures to ensure data preservation.
Confidential
Sr Data warehouse Developer
Environment: UNIX, Oracle 11g, TeradataV13.10, Teradata SQL Assistant and Informatica Power Center 9.1
Responsibilities:
- Created data lineage document (source to Target mapping) based on the load control rules framed by business users.
- Prepare HLD, LLD with data flow diagrams and business requirements. Design walk through with the internal and external stake holders of the project.
- Used Informatica Power Center 9.1 for extraction, transformation and load (ETL) of data in the data warehouse.
- Load and transform large sets of structured, semi structured and unstructured data
- Extensively used Transformations like Source Qualifier, Filter, Router, Aggregator, Sorter, Joiner, Expression and Lookup, Update strategy and Sequence generator and SQL Transformations to perform load control checks.
- Developed complex mappings in Informatica to apply all load control checks for effective pre and post processing of ETL and build centralized data model to store control pass/fail results for all retail loan applications.
- Implemented performance tuning logic on targets, sources, mappings, sessions to provide maximum efficiency and performance.
- Parameterized the mappings and increased the re-usability.
- Used Informatica Power Center Workflow manager to create sessions, workflows and batches to run with the logic embedded in the mappings.
- Provided inputs for the most optimum and feasible design in terms of performance and maintenance after thorough analysis of the requirement
- Scheduled the jobs in Autosys and did monitoring its daily run in Production. Maintain the applications till warranty.
- Analyzed production support related issues, inquiries and requests and provided a concrete explanation of resolutions to issues, inquiries and requests.
Confidential
Data warehouse Developer
Environment: Oracle 11g, Informatica 8.6.1, UNIX, Autosys, TeradataV12, Teradata SQL Assistant and Mainframe JCL
Responsibilities:
- Daily interaction with clients and Business Analysts for requirements gathering, clarifications and suggestions in data modeling and application design.
- Created data lineage document (source to Target mapping) based on the requirements framed by business users.
- Used Teradata for data extraction, transformation and load (ETL) of data in the data warehouse.
- Extensively used Teradata SQL and its OLAP functions such as RANK, ROW NUMBER, CSUM and MSUM etc to perform ETL.
- Developed complex mappings to capture customer data from various sources which lies within enterprise data warehouse to capture customers based on certain parameters and segregate them based on the needs of business analytic users.
- Implemented performance tuning everywhere to ensure ETL runs smooth with increased customer volume day by day.
- Parameterized the components and increased the readability and re-usability.
- Used Mainframe JCL to create jobs, control cards to run Teradata SQL.
- Code Walkthrough and Implement corrections and suggestions before migrating to higher environments.
- Build control plan document for every enhancement for better project tracking and to ensure error free installs and application turnover. Provided inputs for the most optimum and feasible design in terms of performance and maintenance after thorough analysis of the requirement.
- Scheduled the jobs in CA7 and did monitoring its daily run in Production. Maintain the applications till warranty.
- Analyzed production support related issues, inquiries and requests and Provided a concrete explanation of resolutions to issues, inquiries and requests
Confidential
Application Developer
Environment: TeradataV12, Informatica 8.1, Oracle 10g, Teradata SQL Assistant, UNIX and Mainframe JCL.
Responsibilities:
- Mapping Data Items from Source Systems to the Target System.
- Preparing HLD, LLD based on Business requirements preparing Test plan, Unit Test Cases per each requirement in BRD and coordinating with respective source teams for fixing SIT and UAT Defects.
- Used Teradata to create mappings using transformations to move data to a data Warehouse.
- Involved in requirements gathering, creating technical specifications document, developing mappings, testing the mappings, providing support.
- Extensively used Teradata Fast Load, MultiLoad and T-Pump to transfer large volumes of data to BDW model.
- Interact with all the stakeholders for requirements gathering, clarifications and suggestions in data lineage and application design.
- Predominantly focused on building a warehouse model that can fit into bank standards and requirements. Remediate all the queries to ensure performance is improved. Avoid usual problems in data warehousing environment like redundant data across the systems in warehouse, Inconsistent data (not normalized) and occupy huge space etc.
- Scheduling the jobs in CA7 & Autosys and monitoring its daily run in Production.
- Implementing the components and maintain the application till warranty.
Confidential
Application Developer
Environment: TeradataV2R5, Informatica 7.1/8.1, Oracle 9i/10g, Teradata SQL Assistant, UNIX and Mainframe JCL.
Responsibilities:
- Understanding Confidential India reward requirements (HLD/LLD) and preparing all the necessary documents like run book, Low Level Design and DTS.
- Developed reward programs in mainframe Teradata and JCL for Confidential India credit cards.
- Preparing Unit Test Cases and Unit Test Results. Coordinating with respective source teams for fixing SIT and UAT Defects.
- Scheduling the jobs in CA7 and monitoring its daily run in Production.
- Implementing components through changeman package and resolve if any issues post deployment.
- Mapping Data Items from Source Systems to the Target System.
- Preparing HLD, LLD based on Business requirements preparing Test plan, Unit Test Cases per each requirement in BRD and coordinating with respective source teams for fixing SIT and UAT Defects.
- Used Teradata to create mappings using transformations to move data to a data Warehouse.
- Involved in requirements gathering, creating technical specifications document, developing mappings, testing the mappings, providing support.
- Extensively used Teradata Fast Load, MultiLoad and T-Pump to transfer large volumes of data to BDW model.
- Interact with all the stakeholders for requirements gathering, clarifications and suggestions in data lineage and application design.
- Parameterized the components and increased the readability and re-usability.
- Used Mainframe JCL to create jobs, control cards to run Teradata SQL.
- Code Walkthrough and Implement corrections and suggestions before migrating to higher environments.
- Build control plan document for every enhancement for better project tracking and to ensure error free installs and application turnover. Provided inputs for the most optimum and feasible design in terms of performance and maintenance after thorough analysis of the requirement.