Data Architect Resume
SUMMARY
- 14 + years of experience in IT, focusing on ETL, Business Intelligence and Big Data management.
- Extensive experience in ETL & BI Tools: IICS, Power center, DataStage, Talend, SSIS / OBIEE and Tableau.
- Self - starter, motivator, and coach who combines business acumen, corporate culture, organizational goals with the company’s talent pool to maximize their potential.
- Proficient with MPP architecture and large volume data processing.
- Comprehensive knowledge of Business Intelligence, Big Data, Cloud products including emerging technologies as AWS, Snowflake, Google Cloud, Spark, Scala, Hive, NOSQL Dbs.
- Involved in Oracle to Snowflake migration projects.
- Extensively involved in writing Snowsql, creating stages in AWS EC2 instances.
- Extensively worked on Informatica Power Center 10.2/9.1/8.5/7 x. Expertise in Informatica Intelligent Cloud Services (IICS) Data Synchronization, Data Replication, Contact Validation and Data Masking.
- Strong Shell scripting experience in Korn and Bash shells with implementation of dynamic multi-threading and parallelization techniques
- Hands on intelligent structure model, Hierarchy Parser, Data Masking, Web Services transformation etc. and components as Processes, Service Connectors etc.
- Established Jenkins Pipeline to perform continuous Integration in Informatica Intelligent Cloud Services
- Experience in designing REST API's using Python and MuleSoft. Hands on Python scripting.
- Strong knowledge in microservice architecture and RESTful services.
- Skilled at designing and developing high-quality data pipelines, as well as delivering BI solutions through relational and dimensional datasets, data warehouses, and ETL techniques.
- Manage & Perform Analysis, Design, Development, Testing, and Implementation of Business Intelligence (BI) Reporting Solutions in Tableau and hands on Tableau Multi-node Architecture.
- Collaborated with business users in identifying the critical KPIs and metrics, understand the need for these KPIs and reportable metrics.
- Designed logical and physical data models support the analytical needs of the business and interfaces well with BI Reporting tools such as OBIEE and Tableau.
- Performed script execution, maintaining, and building new and/or modifying existing programs to meet system and design specifications.
- Experience in aggregating data, exploring data, building & validating predictive models, and deploying completed models. Preferably using Python libraries such as SKLearn, Pandas.
- Experience in driving the client organization towards effectively making data-driven decisions using data science methodologies like machine learning, effective data visualization etc.
- Strong experience with Informatica tools - Mapping designer, Mapplet Designer, Transformation Developer, Informatica Repository Manager, Workflow Manager and Workflow Monitor.
- Solid Experience in Data Modeling using ERWIN
- Experienced on various databases like Oracle, MS-SQL Server, DB2, Teradata, Netezza etc .
- Experienced with Oracle EBS Reporting, Planning, OneStream, Hyperion and Salesforce .
- Worked extensively with Informatica Designer to create mappings using Expression, Router, Lookup connected and Unconnected, Source Qualifier, Aggregator, and other transformations.
- Experience in full life cycle implementations in Data Warehouse area from designing Logical and Physical Data Models involving Star and Snowflake schemas.
- Understanding and working knowledge of Informatica CDC (Change Data Capture) Implementation experience of CDC using stored procedure, triggers.
- Database Design and Development using Oracle, Teradata, and SQL Server on windows and UNIX platforms.
- Worked in complete Software Development Life Cycle (SDLC) Implementation from requirement gathering, analysis, data modeling, design, testing, debugging, implementation, post implementation support and maintenance.
- Expertise in implementing complex business rules by creating re-usable transformations, and Mappings/Mapplets in Power Center.
- Expertise in importing data from COBOL Files, EDI, Flat Files (Fixed Width and Delimited), XML files and different relational sources like Oracle, DB2, and SQL Server.
- Strong experience in SQL and PL/SQL programming.
- 4+ Years of experience in working with Data Lake environment .
- Experience in creating Transformations and Mappings using Informatica Designer and processing tasks using Workflow Manager to move data from multiple sources into targets and Data marts and Data warehouse.
- Expertise in implementing performance-tuning techniques on Targets, Sources, Mappings & Sessions.
- Expertise in Analysis and E-R modeling, developing Database Schemas like Star schema and Snowflake schema used in relational, dimensional, and multidimensional modeling.
- Have strong RDBMS Concepts and hands on SQL and PL/SQL
- Experience with Shell scripting.
- Experience with Type 1, Type 2, Type 3 Dimensions
- Experience developing Complex mapping using various transformations available in Mapping Designer also in setting up groups, users, permission for Informatica user in Repository Manager.
- Experience with Teradata as Target for the data marts. Worked with BTEQ, Fast Load, MultiLoad
- Proven expertise in analyzing the Business Intelligence needs of the enterprise and building the enterprise Reporting architecture.
- Experience with slowly changing dimension methodology and slowly growing targets methodologies.
- Have good problem solving, time management skills, communication skills, good organizational skills, self-motivated, hard working, ability to work independently or cooperatively in a team and eager to learn.
TECHNICAL SKILLS
Languages: SQL, PL/SQL, Unix, XHTML, Shell Scripting, Perl Script
RDBMS: Oracle 11g/10g/9i, Teradata, DB2, MS SQL Server 2000/2005, MS Access, MS Excel
Big Data Technologies: Apache Hadoop; Hadoop Clusters, MapR 5.5,Cloudera, Hadoop Common, Hadoop Distributed File System; Google Cloud Platform, Replication; Cloudera Cluster; Scala, Kafka, Strom, spark, Big Query,AWS Redshift, AWS Glue, Apache Airflow, Flume, Relational, hierarchical and graph databases, Python, Hive, distributed data file systems, data federation and query optimization
ETL/Reporting Data Integration Tools: IICS (Data Integration, Application Integration, Application integration Console and API Manager.) Informatica Power Center / Power Mart 9.x/8.x/7.x (Designer, Workflow Manager, Workflow Monitor, Repository Manager, Repository Server Administration Console), DataStage, Power Connect for ERP and Mainframes, Power Exchange, OBIEE, Oracle EBS
Operating Systems: Windows Server 2003, Windows 2000, Windows NT 4.0, Windows 98/2000/XP/2007. DOS.
IDE: Toad, SQL*Loader, SQL Server Management Studio, Teradata SQL Assistant, Quest Central for DB2, Change Man Version Manager, Eclipse, Visual Studio, Dreamweaver, Apache Directory Studio, Erwin 4.5/4.0.
PROFESSIONAL EXPERIENCE
Confidential
Data Architect
Responsibilities:
- Design and developing high-quality data pipelines for delivering BI solutions through relational and dimensional datasets, data warehouses, and ETL techniques.
- Create Semantic layer, load legacy data to Snowflake warehouses, scale snowflake warehouses, using lateral flatten functions, using snow-pipes, capacity planning, Monitoring Warehouse Load.
- Collaborate with business users in identifying the critical KPIs and metrics, understand the need for these KPIs and reportable metrics.
- Involved in migrating Oracle DB to snowflake instances.
- Extensively involved in managing snowflake costs, Tuning Snowflake Queries
- Worked on agile methodology in the project.
- Created different transformations such as Joiner Transformations, Look-up Transformations, Rank Transformations, Expressions, Aggregators and Sequence Generator, Sorter Transformation, java transformation, Router transformation, and aggregator transformation.
- Created Batch audit process through command Task in Windows environment as well File archiving, Moving and copying process(Fully Automated Process)
Confidential
Lead Software Engineer
Responsibilities:
- Developed complex mapping also involving SCD Type-I, Type-II, mapping in Informatica to load the data from various sources
- Moved on Perm Application to GCP Cloud using GCP Composer, apache airflow and DAG.
- Performed Bi-Directional Real Time Integration using Informatica Intelligent Cloud Services (IICS) for SAP, Salesforce, ServiceNow and Workday .
- Performed event base integrations using Kafka.
- Designed interoperable machine to machine interaction using Web Service Transformation in IICS.
- Hands on with cloud mapping designer, data masking, data replication, Macro Expressions and Dynamic Linking.
- Experience with REST web services, Application Integration, Application integration Console and API Manager.
- Designed High Level and low-level documents and designed Data Model by Using Erwin.
- Developed SQL Server Integration Services (SSIS) and stored procedures.
- Created uniform interface API's which are unchanging, standardized in communicating between the client and the server using HTTP with URI resources, CRUD (Create, Read, Update, Delete) and JSON.
- Maintained active involvement with the replacement of ETL jobs with SQOOP jobs and HIVE queries including hive performance tuning, and advanced python data models evaluation.
- Developed and maintained complex SQL queries used to create data marts in Teradata .
- Created Metadata Business Intelligence/Tableau framework / views / DataMart’s.
- Use Informatica Designer to create and manipulate source and target definitions, mappings, mapplets, transformations, re-usable transformations, Pushdown Optimization, Batch Audit Process, Fast Load, M Load, Integrated data from the various source system Such as CSV files, MS SQL Server, Teradata, Oracle etc.
- Complete all tasks related to technical analysis, building and unit testing, quality assurance, system test and implementation in accordance with the Technology development life cycle.
- Developed complex mapping to meet customer desired result (Data Integration work) using IICS to load the data from various sources.
- Worked on developing multiple BTEQ Scripts to handle the data conversion and manipulation processes.
- Made use of Collect Stats over the primary indexed columns of the table, defined Secondary Indexes ( USI, NUSI ), Partition Primary Index ( PPI ) and created multiple Volatile and global temporary tables to handle performance of the data load/process and used EXPLAIN PLAN feature to document the load plan/strategy for the development and business teams.
- Used the components of Informatica PowerCenter Designer suite to create multiple data mapping scenarios changing/manipulating the sourced data.
- Created Batch audit process through command Task in Windows environment as well File archiving, Moving and copying process (Fully Automated Process).
- Involved in understanding requirements and in modelling activities of the attributes identified from different source systems, which are in Oracle and CSV FILES. Worked on development of Dynamic Parameter files to capture the changes dynamically and supply the same to respective workflows.
- Helped testing team to develop test scripts as well as did Dry Run Test and Regression testing as well.
- Design parameter file after crating mapping and worked on scheduling jobs using Tidal.
- Created, modified and formatted requests, along with charts and filters, in OBIEE. Built new reports and Customized OOTB reports as per business needs
- Worked for the migration of Repository & Web Catalogs from Development to Test & Production server.
Confidential
ETL /BI Consultant
Responsibilities:
- Perform detailed data analysis, developing online analyst processing (OLAP) cubes for predictive analysis and forecasting.
- Perform File level validations on Input batch file.
- Designed summary report and detailed error report.
- Developed complex mapping to meet customer desired result (Data Integration work), developing high-quality data pipelines
- Load the necessary reference data in data mart. Designed the threshold as per the NCCT business team.
- Designed the ETL to persist all the rejected cost detail records with multiple error codes.
- Created stored procedures, tables, views, synonyms, and test data in Oracle.
- Validate the data in warehouse and data marts after loading process balancing with source data.
- Involved in development and maintenance of the data warehouse for reporting.
- Created effective Test Cases and Integration Testing to ensure the successful execution of data loading process.
Confidential
ETL/BI Consultant
Responsibilities:
- Participated in the Design meetings and created High Level design document as well as low level design document and Data Model by Using Erwin.
- Extensively used Informatica Designer to create and manipulate source and target definitions, mappings, mapplets, transformations, re-usable transformations, Pushdown Optimization, Batch Audit Process, Fast Load, M Load, Integrated data from the various source system Such as CSV files, MS SQL Server, Teradata, Oracle etc.
- Worked on agile methodology in the project.
- Created different transformations such as Joiner Transformations, Look-up Transformations, Rank Transformations, Expressions, Aggregators and Sequence Generator, Sorter Transformation, java transformation, Router transformation, and aggregator transformation.
- Created Batch audit process through command Task in Windows environment as well File archiving, Moving and copying process(Fully Automated Process)
- Developed complex mapping to meet customer desired result(Data Integration work), mapping in Informatica to load the data from various sources
- Collaborated with BI System Analyst to create required reports using reporting tools like OBIEE, SAP Crystal Reports and BI Publisher; profile data necessary to meet reporting requirements.
Confidential
ETL Technical Lead
Responsibilities:
- Provide day to day direction to the project team and regular project status to the customer.
- Participate in the Development and Unit Testing Phases of the Foundation and DD phases of the Project along with the team.
- Handled successful closing of Foundation defects in Foundation layer SIT.
- Conduct Review/Peer Reviews in the development cycle.
- Prepared unit test documents based on each layer's etl logic.
- Provide Technical Guidance to the new joiners/peers in the project.
- Responsible for Work Allocation and tracking the deliverables status.
- Designed Multi day load implementation in ETL.
Confidential
ETL Technical Lead
Responsibilities:
- Responsibilities include designing the documents which is based on the requirement specifications.
- Understand the Business point of view to implement coding using Informatica power center designer.
- Implemented Performance tuning to increase the through put for both mapping and session level and SQL Queries Optimization.
- Automated the jobs thru scheduling using Autosys scheduler, which runs every day by maintaining the data validations
- Provided support and quality validation thru test cases for all stages Unit and Integration testing.
- Prepared/Reviewed project schedules, and design documents developed by Project Development Team.
- Reviewed software deliverables developed by team members.
- Developing test plans, test cases and execute the tests to identify and analyze system Failures.
- Participating in system and integration test.
- Optimizing Query performance and Session performance.
