Etl /bi Developer Resume
Pleasanton, CA
SUMMARY:
- 10 years of IT experience as Data warehouse developer involving in various Full Project management Life Cycles, starting from developing business cases, collecting business requirements, Design Analysis, Development, Testing, Implementation and Maintenance.
- Experience in working on various domains including Finance, Banking, Health - Care, Oil-Gas HR, Insurance and Technology.
- 8 years of in-depth experience in Extraction Transformation and Loading (ETL) processes using DataStage 11.5/9.1/8.7/7.5 Parallel & Server versions and SQL Server SSIS packages.
- Expertise in leveraging wide range of Extraction Transformation and Load (ETL) Tools (DataStage, Informatica and SSIS) for designing and development of Data Marts /Data warehouses.
- 2 years of experience in server less architecture in AWS/Azure in building, storing and processing data through AWS/Azure components.
- Functional knowledge in Hadoop Big data concepts and experience in HDFS, Pig, Hive, HBase, ZooKeeper, Sqoop and Spark.
- Experience in Azure technologies like Azure Data Factory (ADF), Data bricks creating pipeline and activities using both Azure and On-Prem data stores for data loads into a Cloud DW.
- Understanding of when to use Azure SQL DW vs. Azure SQL Server/DB and loading patterns to move data from blob or Azure Data Lake into Azure SQL DW.
- Experience in managing Azure Data Lakes and Data Lake Analytics and an understanding of how to integrate with other Azure Services.
- Experience in Design EDW Data Models and make optimization/improvement recommendations.
- Expertise in writing UNIX shell scripts for encryption (.txt to gpg and pgp), decryption (.asc to .txt files) processes and load the files to vendor site by scheduling the scripts to run on daily basis.
- Expertise in leveraging data warehousing techniques like Change Data Capture (CDC) and Slowly Changing Dimension (SCD) to perform Initial/Incremental loading of Data Marts/Data warehouses.
- Experience in Extracting, Transforming and Loading large volumes of Data from heterogeneous source systems to target systems (Complex flat files, Excel, Oracle, SQL Server, DB2, and Teradata) by performing source, target analysis and developing a Source-Target Mapping Document.
- Experience in ER modeling containing attributes, which can be properties of either the entities or the relationships & Dimensional modeling (Star and Snow flake Schemas) using CA Erwin modeling tool suite to support the Enterprise Information Management initiative.
- Experience with Azure SQL DW and other Azure Services pricing models. Understanding of when to use Azure SQL DW vs. Azure SQL Server/DB and loading patterns to move data from blob or Azure Data lake into Azure SQL DW
- Experience in cloud-based tool ( Power BI) and non-cloud-based tools (Tableau, Microstrategy) for Data Visualization and reporting using KPI’s and the metrics.
- Self-driven, Highly motivated team player with Excellent analytical, problem-solving, documentation, communication, and interpersonal skills.
TECHNICAL SKILLS:
ETL Tools: DataStage 11.5/9.1/8.7/7.5 , SSIS 2008, Quality Stage 8.1.
Databases & Tools: Oracle (11g/10g/9i/8i), SQL Server (2000/05/08), Informix5.1, Teradata v12, JDE, SAP Idoc/ BW/ R/3
Languages: SQL, C, PL/SQL, HTML, XML, MS-Visual Basic, IIS.
Operating Systems: Linux 6.4, IBM AIX (4.0/5.x), Windows XP/NT/2000/7/8.
Test Management Tools: HP ALM/Quality center (11), SAP Solution Manager.
Scheduling: DataStage Director, Crontab Scheduler, Mainframe scheduler (C7), Tivoli Workload Scheduler, AutoSys.
Scripting: Unix Shell Script, Perl, Python, VBScript.
BI Reporting Tools: Power BI,Tableau 10,Cognos 8.3,BusinessObjects XI R2/R3,SSRS.
Testing Levels: Unit Testing, Integration Testing, System Testing, User Acceptance Testing, Regression Testing.
Testing Documents: Test Plan, Traceability Matrix, Test Cases Designing, Test Cases signoff, Test Execution Result log, Defect reporting/Test log, Defect Status Reports, Test Summary Report.
PROFESSIONAL EXPERIENCE:
Confidential, Pleasanton, CA
ETL /BI Developer
- Providing single source to generate, update and manage offers by automating workflow to generate, update and manage Deal Match, Store Coupons, Manufacturer offers, personalized deals.
- Defining and automate business rules to generate, promote and deploy offer definitions and allocations to Just4U.
- Providing single repository to manage all offers. Create extendable offer data model to allow for flexibility to add new offer definitions as needed.
- Creating a reusable process for integrating additional offer aggregators.
- Building process according to Safeway and industry established standards so that the application is simplified, robust, supportable and extendable.
- Meeting all performance requirements as per the business requirements. Making offers available in the Offer Management system for workflow and lifecycle management for all the three offer types by loading separate independent processes.
- Loading Store coupons via an ETL which reads a Spreadsheet provided by Shopper Marketing / Loyalty Initiatives personnel. Manufacturer coupons are loaded via a batch job that fetches offers from the aggregator via their web service.
- Personalized deals are loaded via the UI that is now under the OMS platform and a batch job fetches the offers from TOGM to OMS database
- Managing Offers to go through some level of review by the users of the OMS system before they can be ultimately deployed to production by editing fields, adding comments, and offer lifecycle/workflow management.
- Creating the offer in the J4U system and allow the offer to be previewed in the J4U runtime environment as it will appear on the live site. Setting up offers in Copient prior to load into OMS.
- Promoting offers to production is a signal from OMS to J4U that the offer can be made available on the live site. For manufacturer offers this use case will create the offer in Copient as well. This is step significant because this is the first time the offer is made available to the public to clip.
- Update Offer by making changes to the offer in the OMS system need to be pushed on to J4U if the offer has already been created in the preview step.
- Deactivating Offer can be initiated either from aggregator or from the end user of OMS.
Environment: Linux, DataStage 11.5, DB2 visualizer, Teradata 14.10, WINSCP,Oracle Sql Developer, Shell scripting, Notepad++, Azure Data Factory, Azure SQL DW,Azure SQL Server
UC Office of the President, Oakland, CA
ETL Developer / Business Intelligence Developer
- Straddling the worlds of business and information technology by having a firm grasp of each and being able to mine and analyze data to recommend growth strategies for a company.
- Building solutions to support Data Reporting to meet those business requirements.
- Creating and loading AWS S3 buckets with the data from mainframes for Denodo Database then Extract, Transform and Load data using the AWS ETL Glue and finally loading the AWS Oracle EC2 database using the Informatica, IBM Datastage.
- Writing python scripts for playbooks like creating the database, tables, etc.
- Supporting the Operational Reporting Team in the production of various internal reports.
- Developing, implementing and optimizing data management processes to design reporting and analytics solutions. Identify opportunities to phase out dependencies on legacy databases, processes and reporting tools.
- Working cross-departmentally, to gather business requirements and develop schedules to ensure all deadlines and deliverables are met. Monitor and manage business system data interfaces, files, integrity and security. Assist with design, testing and maintenance of data warehouse. Provide training and end-user support for customized report.
- Designed Functional and Technical specification documents for File Decryption and Encryption Scripting which should be compatible to DataStage ETL run scripts already in PROD environment.
- Created PGP key ring on QA environment, generating private and public keys and performed Unit testing end to end functionality of file encryption and decryption process using user account.
- Coming up with an idea of Creating Common Decryption, Encryption and Archive scripts to be leveraged by all the portfolios by just changing the Prams files in specific to the interface as SPOC.
Environment: Shell Scripting, Python Scripting, DataStatge 11.5, Power BI, WINSCP, Putty, Shell scripting, Notepad++, AWS: S3, Lambda, Kinesis, Redshift, Glue, Quick sight, Cloudwatch, Athena, Tableau v 10, EC2, EMR, DB2, Netezza, UNIX shell scripting (awk, sed),Flat files.
Confidential . Pleasanton, CA
DataStage Developer
- Designing Functional and Technical specification documents according to enterprise standards.
- Requirements gathering from various acquisitions recently done by Confidential ’s Group.
- Functionality analysis for Implementing Slowly Changing Dimensions.
- Analyzing the merger data with main stream data (Safeway way Data) residing on the vendor cloud application.Finalysing the requirements, identifying the hierarchy compatibilities between the groups/categories of data for generating marketing analytics reports.
- Analyzing the Source Systems (both file and tables) then specifying the table structure (metadata) to DB Admin to create tables in Target System.
- Implemented Slowly Changing Dimensions (SCD Type 2) for Store and PMB Clusters.
- Used several stages like Sequential file, Aggregator, Funnel, Change capture, Transformer, Merge, Join, Lookup during the development process of the DataStage jobs.
- Involving in Performance Tuning of various DataStage Jobs/sequencers that are used to load Full extract and the Incremental load and Teradata SQL queries for FastLoad and FastExport scripts.
- Used DataStage Director for running, monitoring & scheduling the jobs.
Environment: DataStage 8.7, Teradata 15.10, SQL Assistant, Axway (Gateway Interchange) UNIX, Putty, MIM Version 8.5.4, OIAM, BTEQ scripts.
Confidential, San Ramon, CA
Sr. ETL Consultant / Business Intelligence Developer
- Worked on sizing each task in Rally and prioritizing the tasks
- Closely worked with Business users to build the product features and functions to fit into the product portfolio.
- Transformed and enriched Data logic is driven by the product features
- Created table Structures in Athena using glue catalog to dynamically support the schema design to enable the business users to query on real time data
- Worked in Athena and Redshift
- Collaborated with the team of twelve developers to identify and communicate value propositions for features and functionality.
- Make trade off decisions in coordination with other product teams to support responsibilities across shared teams
- Work with teams to continuously improve product ownership and other agile processes
- Perform risk management to minimize project risks
- Provide sprint goal and tactical objectives to scrum team, and provide quick JIT decisions to system level questions and defect triage
- Work with program management to track schedule against the roadmap and escalate risks and issues.
- Worked on SOX 404 controls, Internal Controls to help internal auditing process systems.
Environment: S3, Lambda, Kinesis, Redshift, Glue, Quick sight, Cloudwatch, Athena, EC2, EMR, DB2, Netezza, UNIX shell scripting (awk, sed), Flat files, Tidal, Microstrategy
Confidential, CA
Sr. ETL Developer
- Interacting with End-user, IT partner and Business Analyst for defining Business processes Modeling (BPM) document to Improve process efficiency with correct and complete application requirements which will be the feed for creating a functional specification document.
- Develop Functional Specification Document (FSD) which comprises of Business Requirements, Expectations, Technology, IT Infrastructure Requirements, Data Security and Privacy Requirements.
- Expertise in Performance tuning of DataStage jobs by leveraging inbuilt functionalities like the job monitor to get CPU summary information where time is being spent in the flow, Usage of Score Dump, Resource Estimation assist and monitoring performance statistics.
- Designed Master Job Sequencers which are triggered on daily, weekly and monthly basis from crontab scheduler to deliver the encrypted files to the vendor site SAP- Success Factors
- Co-ordinate the defect review process with the vendor (SAP- Success Factors ) and team members to monitor, and report progress during each test phase or cycle.
Environment: InfoSphere DataStage 8.1/8.7, FTP Server, Excel, Cron Shell Scripting, Sql Server 2008. Oracle Database 11g, Red Hat Enterprise Linux Server 6.4, HP ALM , SAP- Success Factors, SAP BW v.7
Confidential, CA
Sr. ETL Developer
- Analyzed the database schema structures of the OLTP databases and designed ETL processes.
- Worked closely with Development, QA and Business groups to understand the data needs for the creation of reusable data sets to support Development and Testing Teams
- Used Agile SCRUM methodology and RallyDev for bi-weekly sprints to track the progress and effort estimation for user stories.
- Created Parameters, Environment variables, Stage Variables and Routines to run the same job for different schemas and debugging using Peek, Head and stages
- Experience with UNIX shell scripting for checking the data and number of records loaded into the table on day to day basis.
- Perform development in development environment and resolve performance Issues, test in staging environment and work with other groups to push changes to production
Environment: DataStage 8.7 Parallel, Ascential DataStage, SAP BW,SQL server 2008,Oracle11g, PL/SQL, and Information Analyzer 8.1, Rallydev, SAP CRM,SAP IDocs, SAP R/3.
Confidential, TX
Sr. ETL Developer
- Analyzed Business Requirements by working closely with Business Analysts.
- Created DataStage jobs and sequencers to extract data from various sources extracting the data from SAP R3, flat files, SQL Server, Oracle, transforming data according to the business requirement and loading into target Oracle tables.
- Data Profiling the source and Target systems to identify potential data issues
- Used DataStage Director to validate and troubleshoot the jobs. Participated in weekly status meetings.
Environment: DataStage 8.7 (server and Parallel), Showcase V11, Oracle11g, PL/SQL, SQL server 2008, SSIS 2008, Information Analyzer 8.1, HP ALM, SAP R3
Confidential, CA
Sr. ETL Developer
- Participate in Project scoping activities and Understanding requirements as per the Client Business Logic’s.
- Preparing Information Analysis docs and Low Level Design Docs for ETL Process.
- Configure Database connectivity and performing elementary Data Stage administration activities.
- Involved in development phase meetings with Business Analysis and end users for requirements Gathering which is a lead to Data Analysis docs and Low Level Design Docs for ETL Process.
- Developed customized routines, wrote stored procedures, and created functions and packages
- Ensured acceptable performance of the data warehouse processes by monitoring, researching and identifying the root causes of bottlenecks.
Environment: DataStage 8.5 (Parallel), Oracle11g, PL/SQL, Unix/Window, MS Access 2010, SQL server 2008 R2, UNIX Shell scripting, cron Shell Script.
Confidential, MI
Sr. ETL Developer
- Interacted with Business Analysts to finalize the requirements and documented the technical design document for Coding.
- Developed Data Stage Parallel Jobs using required stages, and then obtained data from source like flat files, Cleansed, summarized, aggregated, transformed and loaded into data warehouse.
- Encompasses Database designing, ETL phases of Data Warehousing. This is achieved with an emphasis on relational data modeling and dimensional data modeling for OLTP and OLAP systems.
- Experience in developing complex transformations by leveraging the inbuilt functionalities.
- Data staging of the ETL solutions from different source systems to target database.
- Involved in the identification of Facts and dimensions table for developing logical design of the data marts.
Environment: InfoSphere DataStage 8.5 (Server & parallel), Ab Initio 2.14,Teradata v12, PeopleSoft EPM 8.9, UNIX- AIX, Windows XP, MS Word, Excel, HP ALM, Korn Shell Script.
Confidential
Sr. ETL Developer
- Understood functional documents for business source systems and applications and their integration into the DWH.
- Wrote the Design documents and specifications in accordance with SAMS (Statewide Accounting Management Systems) manual design and best practices.
- Involved in understanding the Requirements of the end Users/Business Analysts and Developed Strategies for ETL processes.
- Data Analysis using IBM Info Sphere Information Analyzer, Configuring resources, Importing metadata, Specifying analysis settings, Analyzing columns, Publishing analysis results, Analyzing tables, Analyzing across tables, Managing metadata.
Environment: DataStage 8.1,Oracle11g,PL/SQL, Windows 98,MS Access 2010,PeopleSoft EPM 8.9, SQL server 2008 R2, Unix shell scripting.
Confidential
Sr. ETL Developer
- Interacted with business users in gathering the business requirements, identifying the relationship within and between the groups of data.
- Created the Functional/Technical specification documents by maintaining standards of the enterprise.
- Designed the technical document for various jobs, unit test documents with various scenarios and acceptance level of test case results.
- Expertise in writing Unix shell scripts to encrypt, decrypt files to perform sFTP of files over outside servers by triggering the datstage master sequencers jobs from cron scheduler and sending out the mail alerts whenever there are failure scenarios for error Handling and reject management, Audit management process.
- Involved in various Testing Procedures such as Source to Target testing (validation testing), Application Upgrades testing, Metadata testing, data Completeness testing, data Accuracy testing and finally Production data validation testing.
Environment: DataStage 8.1 Enterprise Edition, Teradata v12,Oracle 10g,TOAD 7.4, PL/SQL, Unix Shell Scripting, Windows NT server, AIX UNIX 5.2.
Confidential
Sr. ETL Developer
- Involved in requirement gathering and performed data profiling by analyzing source and the target systems and thereby creating a Source - Target mapping Document.
- Developed subject matter expertise with source and target systems by creating Source-Target data mapping document and using DataStage Designer for extracting data, performing data transformations
- Used Surrogate Key Generator to generate surrogate key columns which is a substitution of natural primary key for a Dimension table and used to join a dimension table to a fact table.
- Used DataStage Director Log to track the defects and issues.
- Analyzed technical specifications and formulated test data strategies by creating test data for comprehensive test coverage in incremental data validation.
- Developed UNIX shell scripts to automate the Data Load processes from source to the target Data mart by scheduling the scripts in crontab on daily, monthly and yearly basis.
Environment: DataStage 7.5 PX, Quality Stage, DB2 Version 8.0, AQT, AIX, SQL,PL/SQL, PeopleSoft EPM 8.9, UNIX Shell Scripting, Tivoli work scheduler.
Confidential
ETL Developer
- Interacted with business users in gathering the business requirements, identifying the relationship within and between the groups of data.
- Created the Functional/Technical specification documents by maintaining standards of the enterprise
- Designed the technical document for various jobs, unit test documents with various scenarios and acceptance level of test case results. Migration of Server jobs from DataStage 7.5 to 8.7 versions.
- Analyzing the showcase query for the sources system JDE and specifying the Metadata to the DB admin to create the tables in the SQL target Database.
- Implementing Type-2 Slowly Changing Dimension for capturing the delta and Using Shared Containers for reusability and thereby reducing the job complexity.
- Provided program documentation, including restart/recovery procedures. Created a flow chart diagram to demonstrate the process to the team.
Environment: IBM Web Sphere DataStage 8.0.1, QualityStage 8.1, SAP R3,Oracle10g, DB2 UDB,SQL Server 2005,Autosys,PL/SQL,UNIX,Business Objects V6.5.
Confidential
ETL Developer
- Involved in reviewing Business/System requirements and other deliverables.
- Reviewed the data models, maps and documents to ensure that the design solution meets the requirements
- Responsible for validating the DataStage jobs, sequences against the pre-defined ETL design standards Tuned Source, Targets, jobs, stages and Sequences for better performance
- Used DataStage Parallel Extender for splitting the data into subsets and to load data, utilized the available processors to achieve job performance, configuration management of system resources.
- Created user defined transforms and routines for handling special cases
- Extensively used SQL and PL/SQL coding in DataStage jobs for processing data.
- Written UNIX shell scripts to validate ETL processes to check the functional and nonfunctional testing requirements.
Environment: DataStage 7.5 X2, Oracle 10g, UNIX, Shell Scripting and AutoSys tool.
Confidential
ETL Developer
- Worked on the logical and physical design of the Data warehouse. Identified sources/targets and analyzed source data for dimensional modeling.
- Good knowledge on credit increase plans to employers to offer new credit card packages. Worked in design of good credit, bad credit and credit increase data marts.
- Used DS Erwin Meta Broker to import Erwin 4.x Metadata into Data Stage Repository.
- Created Business Objects reports, Queries with constant interaction with the end users. Trained end users in understanding the reports. Functionalities such as Slice and Dice, Drill mode and Ranking were used for Multidimensional Formatting.
- Web Intelligence was used to generate reports on the internet/intranet.
- Provided 24/7 production support.
Environment: DataStage 7.5, Meta stage 7.0, Business Objects 6.5, Oracle 9i, PL/SQL, SQL Plus, UNIX Shell Scripts, Windows 2000, and ERWIN 4.1.