We provide IT Staff Augmentation Services!

Data Services Architect Resume

CA

SUMMARY:

  • Experienced SAP HANA, BI and Analytics Consultant, with 6 years SAP HANA/Analytics, 12+ years of SQL, 2 years of Hadoop/Big Data, 9 years of SAP BI, 6 years of ABAP and 8 years of ETL (BODS, MuleSoft), Python for data wrangling.
  • HANA Architecture/Development - Architect and develop attribute, analytical, graphical and scripted calculation views, ETL to HANA from various sources, data transformation inside HANA for data marts using SQL programming
  • ETL - BODS (Data Services), MuleSoft, Informatica, Python - Data transformations from Big data Hadoop/Hive, Kafka, Hana, Sales Force, NetSuite, SAP BW, SAP ECC, MSSQL server, MySQL
  • Experience working with Python for data wrangling (Numpy, Pandas), visualization (Matplotlib, Bokeh) and machine learning (Scikit-Learn)
  • Experience working on AWS EC2 instances for HANA, Hadoop
  • Experience working with MapR and Cloudera Hadoop implementations
  • Big Data - Hive, Impala, Kudu, Sqoop, HBase, HDFS, Hadoop, Kafka, Flume, Spark.
  • Analytics/Front End development - Tableau, Qlik View, BOBJ (WEBI, Explorer, IDT, Relational Universe).
  • Worked on Implementations, upgrades, performance enhancements in the HANA, BI/BW and Analytics space.
  • Extensive experience in data analysis and root cause analysis and proven problem solving and analytical thinking capabilities.
  • Self-motivated team player with a passion for learning new technologies. A steady performer with efficient multi-tasking in high stress or fast-paced environments.
  • Ability to prioritize workload and work with minimal supervision.
  • Excellent communication skills and well versed in gathering requirements from end users.
  • Other experience includes research and University teaching career in Physics (Associate Professor).

SAP HANA/SQL EXPERTISE

  • Architect and develop HANA Information Models - Design & Creation of HANA schemas, HANA Attributes views, HANA Analytic views, HANA Calculation views (both Graphical and Script), Analytic privileges.
  • Built data warehouse on stand-alone HANA with data from various sources like Big Data Hive, NetSuite, SalesForce, Oracle, MySQL, SQL Server, Sybase IQ, SAP ERP data, flat files
  • Configured and used SDA (Smart Data Access) for various remote sources to expose as virtual tables in HANA
  • Design and implement complex transformations in Hana using SQL statements, Stored procedures User defined functions and Triggers. Used SQL Window functions (Rank, Lead, Lag ) extensively
  • Implement and schedule HANA stored procedures for ETL into HANA.
  • HANA Information model tuning using Plan Vizualizer
  • SQL tuning in Stored Procedures and scripted calculation views
  • Used temporal joins when necessary for reporting on SCD (Slowly Changing Dimensions)

ETL EXPERTISE

  • BODS (Business Objects data Services) to load non-SAP data into HANA from a variety of sources such as, Hadoop Hive, SalesForce, NetSuite, Teradata, SQL server, My SQL, Sybase IQ, SQL Anywhere, Oracle. Used Table Compare, History Preservation, Key Generation, Reverse Pivot, Query and SQL transforms
  • SLT replication to load Finance tables into HANA, implementing transformation rules in IUUC REPL CONTENT. Implemented ABAP code/includes in transformation rules to address specific data load business requirements
  • HANA SDA (HANA Smart Data Access - Accessed tables from DB2 as virtual tables and then persisted the data as COLUMN tables in HANA
  • Informatica Cloud to load data from NetSuite to Hana

BIG DATA EXPERTISE

  • Expertise in Hive QL. Created, partitioned and manipulated external tables, views in Hive
  • Used Parquet, ORC, Text formats for HDFS files
  • Sqoop ETL to transfer data from Sales Force, MySQL, SQL server to HDFS/Hadoop, Hive and Impala
  • Used BODS (Business Objects Data Services) to ingest data from Big data Hadoop Hive. Developed parallelization techniques to move big data (4-5 terbayte tables loaded into Hana from Hive, with record sizes of 5 - 10 billion in a matter of hours)
  • Real-time data ingestion using Kafka and Spark Streaming into HBase
  • Updates and Inserts in Big Data in Kudu
  • Data Mart development in Hive/Impala using SQL
  • Analytics in Tableau and QlikView Python, Machine Learning Expertise
  • Tuples, Lists, Dictionaries, Numpy arrays, Pandas (Series and Dataframes) for Data Wrangling
  • Scikit-Learn for Machine Learning sklearn.linear model (Linear, Logistic Regression, Lasso, Ridge Regression), sklearn.tree (Decision Trees), sklearn.svm (linear and radius kernels), sklearn.neighbours (KNN), sklearn.ensemble (Random Forest), sklearn.model selection (train test split, LeaveOneOut and KFold CV, ShuffleSplit) etc
  • Matplotlib, Seaborn, Bokeh for visualization

ANALYTICS EXPERTISE

  • Tableau - Connected to various data sources like HANA, MySQL, SyBase, Hive/Impala, Local Files and created Visualizations, Dashboards.
  • QlikView - Connected to data sources - HANA, MySQL, SyBase, Hive/Impala. Built complex data marts in QlikView Script
  • BOB J - Relational Universe design with table joins and resolution of loops and traps using aliases and contexts. Usage of @Functions. Extensive experience in writing free-hand SQL, stored procedures, user defined functions and complex joins.
  • SAP BEx Experience in working with Query Designer and Analyzer. Designed and created Queries/Reports using conditions, exceptions, restricted/calculated key figures, filters, variables, customer exits, workbooks and jump targets.

SAP BW ON HANA EXPERTISE

  • Involved in classic SAP BW to SAP BW on HANA migration.
  • Existing BW 7.1 stack-split (ABAP and JAVA) performed, upgraded from BW 7.0 to BW 7.3, and migrated Db from Oracle to SAP HANA, OS from Solaris to Linux
  • Implemented PCA for cloning of delta data queues from a single ECC system to both classic BW and BW on HANA systems
  • Involved in optimizing Infocubes and DSOs using TCode RSMIGRHANADB
  • Imported classic BW infocubes into HANA as HANA models for BOBJ query optimization
  • Optimized existing classic BW process chains by removing Attribute Change run, create/delete cube index, BWA rollup etc.
  • Worked on ABAP code optimizations for HANA using ATC (ABAP Test Cockpit) and SQL Monitor

SAP BW EXPERTISE

  • Strong understanding of concepts behind star schema, enhanced star schema, structure of database objects behind the BI objects
  • Worked in Business areas of PM, SD, MM (Inventory Management and Purchasing), FICO.
  • End-to-end BI/BW development - Data Extraction (Generic Datasources and BI Content), Staging, Modeling, Transformation, Data loading and Reporting.
  • Experience in creation and use of Generic, LO Cockpit, CO-PA, LIS, FI-SL datasources. Good knowledge of backend tables and dataflow of PM, SD, MM and FI modules for creation of database views and Function Modules for cross application generic dataSources.
  • Extensive use of ABAP in end routines in Transformations for using Write-Optimized DSO as data stores. Wrote custom ABAP code for identifying delta records on a package by package basis.
  • Designed Fact tables, dimensions, line dimensions, display and navigational attributes and created Objects for Data Loads. InfoObjects, InfoSources, Transfer /Update Rules, Transfer/Update Routines, InfoPackages, Transformations, DTPs, InfoProviders like ODS and DSO, InfoCubes, Multiproviders. Created and monitored Process Chains for the data loads.
  • Monitoring of data loads and troubleshooting of failed data loads
  • Involved in performance tuning at modeling, data load and Query design levels. Used sound, approved design recommendations for dimension, infoCube design, InfoCube Aggregates, Indexes and Partitioning.

SAP ABAP EXPERTISE

  • Experience in enhancement of BI DataSources (CMOD), creation of function modules for generic datasources, ABAP routines for transformation/Update rules and filtering in DTP packages, file selection routines in InfoPackages.
  • ABAP programs for creating flat files for loading Hierarchies and Hierarchy texts
  • ABAP Function modules for Utilities, using field-symbols for dynamic comparison of fields
  • Data Transfer using BDC, Call Transaction for Customer, Material and Vendor Masters, Sales Orders.
  • Simple and Interactive ALV Reporting on backend data from SD and MM modules.
  • User Exits for data extraction, and BEx variables, virtual key figures(CMOD, BADIs).
  • Form printing with SAP Scripts, and SmartForms development.
  • Well versed with DDIC (Data Dictionary), internal table programming.
  • Created database Views and Function Modules for BI Generic DataSources.
  • Dialog programming.

TECHNICAL SKILLS:

HANA: HANA Information Modeling on SAP HANA 2, SPS 12 revision 122, 10.9.0, 7.0, 6.0

Data Warehousing: SAP HANA Standalone Warehouse, SAP BW 7 - 7.3 Migration to HANA, SAP BI (7.3, 7.0, 3.5), Microsoft BI

Big Data: Map-R and Cloudera Hadoop implementations, Hive, Impala, Kudu, Sqoop (ETL), Kafka, Spark

ETL: BODS (Data Services), Python Pandas, MuleSoft, SDA (Smart Data Access) in HANA, SLT, Stored procedures for data transformation in HANA, Informatica, BW data flows

Data Science: Machine Learning Modeling with Scikit-learn

Cloud/Apps: AWS, Salesforce, NetSuite

Databases: Big Data Hive, Impala, HBase, SAP HANA 2.0, Sybase IQ, SQL Anywhere, MySQL, Microsoft SQL Server(2014, 2012, 2010), Teradata, ORACLE 12, 11i, 10g, 9i, 8i, Pervasive, BTrieve

Programming: SQL, SQL Stored procedures, HANA SQL Script, Python, Triggers, SAP ABAP, C#, VB.NET, VB, Java, COM, XML, ASP, JSP, VBScript, HTML, CSS

Analytics: Tableau, BI 4.0 (WEBI, Explorer, Dashboard, Crystal Reports), LUMIRA, BEx Analyzer, BEx Query Designer, Microsoft SQL Server Reporting (SSRS)

SAP Releases: CRM, SAP BW 7.3, SAP ECC 7.0, ECC 6.0, ECC 5.0, R/3 4.7, 4.6c, SAP Solution Manager

Functional: SAP BI/BW, Data Flow and DB tables of CRM, PM, MM, SD and FI Modules, NetSuite Functional/Tables

Platforms: Windows (10, Vista, XP, 2003, 2000, NT, 95/98, Me), Ubuntu, RHEL

Version Control: HANA Repository, GIT, Visual Source Safe, Rational Clear Case, CVS (Subversion, Tortoise)

PROFESSIONAL EXPERIENCE:

Confidential

Data Services Architect

Responsibilities:

  • Built complex models in Hana using SQL stored procedures and SQL script in Hana Calculation views to derive the Installed Product Foundation, to replace the processing that existed in the Teradata legacy system
  • Worked primarily with Install Base and Service Agreements data to derive metrics for cross-sell and upsell opportunities for Cisco products. Replaced a legacy system built in Teradata with a new data warehouse in Hana
  • Record counts for some of the transaction data is in the billions. Used Partitioning of tables (hash) in Hana since record count is over 2 billion for most of the transaction tables. Designed SQL stored procedures to process the data in partitioned chunks.
  • ERP system of Record is Oracle ERP for Confidential, Inc. Transaction history is archived in Hadoop (Map-R distribution) under the Hive database. Hadoop HDFS file size per transaction table on average is about 3-4 terabytes, with some tables having record counts of more than 10 billion
  • Major challenge is to bring this big data into Hana for fast analytics using SAP BODS. Since the Hive connector delivered by SAP BODS is new, but limited in functionality, faced many limitations/challenges in importing data from Hadoop Hive.
  • Architected and developed a combination of parallelization techniques in Data Services, partitioning in Hadoop, partitioning in Hana (Hash), SQL transform to query data from Hive source, to move the data at a fast transfer rate. Scaled up memory on both Data Services and Hadoop to support parallelization. Achieved a fast transfer rate of 18 million records per minute
  • Used Python to validate record counts, distinct values per table column between source Hive and target Hana after the data import

Confidential

HANA Architect/Developer, Big Data

Responsibilities:

  • ERP system of Record for Fitbit is NetSuite. Existing legacy system used ‘saved searches’, which are compiled queries in NetSuite, created by Power users of the Sales and Distribution division. These saved queries were used to generate flat files which are then imported into Microsoft SQL server using SSIS. SSAS cubes built upon the data in SQL server were used in excel reporting.
  • The objective/project was to create a new data warehouse in Hana that would replace the legacy system mentioned above.
  • The ‘Saved Searches’ are not transparent, so a high degree of reverse engineering was necessary to get to the underlying table, columns and correct joins to form the base for the stand-alone warehouse in Hana
  • Architected and developed the Hana stand-alone data warehouse from scratch for the Sales and Distribution division.
  • Used Informatica (existing) and BODS (new implementation) to design and implement the ETL - Data is imported from sources: NetSuite (ERP data)SQL Server (subset of Master data) and Zyme (Channel Data Management Software) flat files from fitbit distributors on inventories procured and sold into: target data warehouse on Hana
  • Architected and developed the complex transformations on this base data in Hana using SQL, stored procedures, Calculation views, to derive the metrics for the Sales and Distribution Division
  • The metrics derived above are then consumed by BOBJ and tableau reports

Confidential

HANA Architect/Developer, Big Data

Responsibilities:

  • Worked primarily for GCS (Global Customer Support) department. This department relies heavily on metrics developed in HANA using CRM data from Sales Force, SAP and MySQL databases
  • Solely responsible for architecting, developing and deploying CRM data marts for all sub departments within GCS
  • ETL - imported CRM data from Sales Force, NetSuite (legacy), MySQL, SAP, Sybase IQ, flat files into HANA using both BODS and Mule Soft
  • Used SDA (Smart Data Access) to integrate data from remote sources into HANA models
  • Extensive use of stored procedures (both read-only and read-write), HANA views (attribute, analytical and Calculation Views)
  • Developed Dashboards for GCS External Partner data, with HANA views as the backend
  • Although Sales Force backend data for date times is in GMT, the BODS driver for Sales Force brings in data in PST (automatic conversion). The GCS department needed the metrics in GMT and the BODS driver had no settable parameter that would bring the data in GMT. Conversion from PST back to GMT results in erroneous data around the Daylight savings time. So, used Mule Soft to connect to Sales Force to bring in both GMT data and PST data (converted in the data transformations, there is no conversion loss from GMT to PST). For this effort needed to drop and reload all Sales Force tables in HANA.
  • Retro-fit all the existing metrics code base (HANA Stored procedures, HANA Views) from PST to GMT
  • Developed the GCS metrics in Big Data, since the new direction from management was to move to Big Data from HANA - built the POC for bringing the CRM data into Hadoop
  • GCS Metrics in Hadoop: Imported data from Sales Force, HANA, MySQL into HDFS/Hive (data warehousing for Big Data) using Sqoop. Exposed data in Impala (fast analytics for Big Data). Re-developed the GCS data marts using SQL in Impala.
  • Analytics of data in Impala in Tableau (also Qlik View)

Confidential

HANA Architect

Responsibilities:

  • Existing system had SAP CRM data, extracted into BW using SAP extractors, fed to Sybase IQ database via Open HUB, further SQL transformation of data in Sybase IQ, to WEBI reporting using Universes. Requirement was to replace the BW system and Sybase IQ system with HANA views directly on SAP CRM data tables replicated into HANA
  • Reverse engineered BW extractors by debugging the ABAP code in the extractors to come up with requirements/design for the HANA information models
  • Replicated SAP CRM tables into HANA using HANA SDA (Smart Data Access) - Exposed DB2 tables as virtual tables in HANA and then persisted the tables in HANA
  • Designed and built HANA information models. Designed and built HANA Attribute, Analytic and Calculation views (both Graphical and Scripted)
  • Created Scripted Calculation views for complex scenarios, used SQL window functions. Created read-only stored procedures for consumption in the scripted Calculation Views.
  • Created read-write Stored Procedures for transforming the CRM tables and persisting data in HANA
  • Used SDA (Smart Data Access) to integrate data from remote sources into HANA models
  • Created Analytic privileges for applying security on the HANA information models. Created Dynamic Analytical Privileges as well using stored procedure for the filter values
  • Fine-tuned HANA information models using Plan Vizualizer for better performance
  • Led and educated a team of 9 developers in HANA information modeling

Confidential, CA

SAP BI HANA Architect/Developer

Responsibilities:

  • Conducted Workshops with Juniper Business Owners in the areas of Asset Accounting, Accounts payable, Accounts Receivable to determine requirements for custom BI reports
  • Conducted Design Briefing sessions with Off-Shore teams and oversaw the Development, Testing and Deployment of the backend objects and custom BI Reports
  • Architected and built the HANA information models/views. Created Attribute, Analytical and Calculation views (both Graphical and Script)
  • Created SQL stored procedures in HANA to address business requirements. Well versed in SQL and CE scripts
  • SLT replication to load Finance tables into HANA, implementing transformation rules in IUUC REPL CONTENT. Implemented ABAP code/includes in event-based rules to address specific data load business requirements
  • Used BODS to load non-SAP data (flat files, from SQL server) into HANA. Used Table Compare, History Preservation, Key Generation, Reverse Pivot, Query and SQL transforms

Confidential, CA

SAP BI Architect

Responsibilities:

  • Architected and Developed Performance Enhancement for existing BEx Reports
  • Implemented SPO (Semantically Partitioned Object), Logical Partitioning, Physical Partitioning, Re-modeling of Dimensions to improve performance
  • Dimension Re-modeling brought down Fact table ratios from greater than 100% to less than 5%
  • Improvement of Query times using Aggregates and Pruning from several minutes to a few seconds

Confidential, CA

SAP BI Architect/Developer, HANA modeler

Responsibilities:

  • HANA Modeling, SQL Stored procedures using SQL script, CE script, Attribute/Analytic/Calculation Views, Analytic Privileges. As part of moving to Enterprise HANA, built HANA information models with SQL scripts/procedures to reproduce/validate reports already existing in classic BW system.
  • Involved in SAP BW on HANA migration. Existing BW 7.0 system stack-split (ABAP / JAVA) performed, upgraded to BW 7.3, and migrated to SAP BW on HANA
  • Implemented PCA for cloning of delta data queues from a single ECC system to both classic BW and BW on HANA systems
  • Involved in optimizing Infocubes and DSOs for HANA using TCode RSMIGRHANADB
  • Imported classic BW infocubes into HANA as HANA models for BOBJ query optimization, optimized application logic by encapsulating business logic in SQL procedures, using SQL scripts only or CE scripts only
  • Optimized existing classic BW process chains by removing Attribute Change run, create/del cube index, BWA rollup etc for the HANA system
  • Re-architected the transformation/data cleansing code by pushing logic into the database layer (SQL stored procedures using SQL Scripts / CE scripts)
  • Sole developer/architect for many finance and legal projects at Confidential .
  • Extremely good understanding of functional and technical data patterns.
  • Active participation in functional requirement gathering for legal project which automated the Employee Stock Purchase Plan. Architected and developed the data flow of stock awards/options and releases/exercises from ETrade application into BW. Built a transactional system where a user can approve tax payments on the sale of releases/exercises to be sent on to Payroll departments all over the world (for the different Confidential centers).
  • Developed the legal solution mentioned above, in the BW on HANA platform. Noted significant improvement in DSO activation and load times (a 1 hour 45 minute load time in classic BW came down to 8 minutes in the HANA system)
  • Designed and developed solutions using non-cumulative key figure architecture for determining account balances as of a given date from bank data incoming in SWIFT files
  • Designed and developed the different data targets (InfoCubes, DSOs, Infosets, MultiProviders, Master data objects) and wrote extensive ABAP code in Expert/Start/End routines.
  • Wrote Function Modules that served as utilities - for example, wrote a Function Module that reads an incoming flat file, validates (character fields) and replaces invalid BW characters with a valid BW character that can be specified as a user specified parameter, generates a cleansed output flat file for BW consumption. This FM can be used generically for any flat file. This FM works by dynamically assigning the columns of the flat file to field symbols.
  • Also wrote Function Modules that serve to get technical information on BW objects (infoObject list for DSO or Infocube along with descriptions etc.) to use in technical documentation for all projects at Confidential .
  • Performance Improvement of data loads by tuning ABAP code in transformations and look-ups.
  • Sole architect/developer for an important project to include new attributes in an existing master data object and restate the master data object that was used pervasively in each and every finance solution at Confidential . Responsible for the ABAP code in this restatement of attributes. Also solely responsible for data validation of all financial solutions in Confidential after the restatement of this attribute. There was not a single production issue with restatement of all this data. Have good understanding of financial data.

Confidential, CA

SAP BI Architect

Responsibilities:

  • Involved in Blue Print Realization for Order-to-Cash and FICO BI reports. Involved in the European (Belgium) implementation.
  • Created Technical Specifications for BI backend objects, Bex Reporting and Business Objects WEBI reports according to Functional Specifications after discussion with the super users of the different functional areas
  • Designed and developed write-optimized DSOs for Staging, Standard DSOs for functional Storage, Transformation including Start, End and Expert routines (ABAP code) for backend BI Objects. Enhanced Master DataSources, Transaction Datasources (CMOD) and created Generic Datasources for AFS (Apparel Footwear Solution) using Function Modules.
  • Performance enhancement of data loads using FAE hints for ABAP code and Secondary indexes for TeraData DSO tables.
  • Designed and built Connections, Data Foundations, Business layers, Universes for Business Objects WEBI reporting using BICS connections for BEx queries and JCO connections for DSO reporting.
  • Collected and transported both ECC and BI developments across the landscape. Involved in FAT, SIT testing and final production validation of the reports.
  • Technical support of the new Reports after the Go-Live.
  • Interacted with and managed Off-Shore development teams, both in the SD and FICO areas.

Confidential, CA

SAP BI Architect

Responsibilities:

  • Involved in Requirements gathering from meetings with business users. The broad requirement is to extract ECC data from the Plant Maintenance Work Order data and Material Management Inventory data (to match MC.9 ECC report) to load into BI, without using LO Cockpit datasources.
  • Created the technical specs from the general, broad business requirements. The technical specifications detailed the list of backend tables, extract structures, function modules for generic datasources, write-optimized data stores, end routines for data staging, infocubes for reporting, BEx query design for KPFs and summary/detail reporting.
  • Designed and created generic datasources by writing function modules to extract Plant Maintenance Work Order detail (Operations, Reservations, Goods Issue and Confirmations).
  • Enhanced master datasources in CMOD for Plant Maintenance Work Order header data.
  • Used write-optimized DSOs as Data Stores. Coded end routines to filter and load delta records into write-optimized DSOs, on a package by package basis. Using Write-optimized DSOs resulted in vast reduction of activation time.
  • Designed and created generic datasources for extracting material inventory data from MARC, MARCH, MARD, MARDH tables for stock quantities and MBEW, MBEWH tables for material pricing. Coded end routines in transformations for explosion of these tables to fill in gaps for fiscal periods.
  • Used 0CO datasources to load data for Work Order Accounting/Costing data for material and labor (value categories).
  • Created BEx queries for Key Performance Indicators, and reports to validate against ECC MC.9 reports.

Hire Now