We provide IT Staff Augmentation Services!

Big Data Architect/scientist Resume

4.00/5 (Submit Your Rating)

Richardson, TX

SUMMARY:

  • Total 15 years experience in handling OBIEE, Tableau, Power BI, Qlikview, Adobe Analytics, Spotfire, Cognos, Oracle, SAS, SPSS, DB2, SQL Server (SSIS, SSRS and SSAS), Teradata and UNIX. 3 years experience as a DBA. Strong background in system design, development and testing.
  • Solid understanding of Data Warehousing, BI life cycle and strong working experience in statistical analysis, quantitative analytics, forecasting/predictive analytics, multivariate testing (A/B testing), and optimization algorithms.
  • Certified Datawarehouse Developer and BI developer.

TECHNICAL SKILLS:

ETL Tools: Informatica, SSIS, Oracle Data Integrator, Cognos Data Manager, Quality Stage.

Big Data: CDH Hadoop/Hive/MapReduce/Pig/Sqoop/Flume, HDFS, Scoop, Apache Spark, Autosys, Oozie.

Machine Learning: SAS, R, Matlab, Scikit, SPSS.

OLAP: Cognos PowerPlay, Cognos Transformer.

Reporting Tools: Qlikview, Tableau, SSRS, SSAS, Power BI, Cognos BI, Microstrategy, OBIEE, Spotfire, Adobe Analytics, Oracle Discoverer.

Data modeling Tools: Erwin, star schema, snowflake schema, Visio.

RDBMS: Oracle, MS SQL Server, Teradata, Sybase, MS Access, DB2, SQL *Plus, SQL Loader

Languages: C, C++, Java, Pl/SQL,Python, Hadoop, MapReduce, Pig, Hive.

Environment: s: SAP, JDE, PeopleSoft, UNIX.

PROFESSIONAL EXPERIENCE:

Confidential, Richardson, TX

Big Data Architect/Scientist

Responsibilities:
  • Expertise in implementing complex business rules by defining output layouts for Sales, Order Management, Provisioning, Billing, Trouble Ticketing and Financial system reports using Tableau,Spotfire, Cognos and Power BI.
  • Creation of data Mart by extracting data from Mainframe, DB2 and SQL server using informatica and SSIS.
  • Experience in statistical analysis, quantitative analytics, forecasting/predictive analytics, multivariate testing (A/B testing), and optimization algorithms.
  • Design and develop Mappings to load data from multiple source system databases (DB2, MAINFRAME, SQL Server and flat files) into the DW.
  • Used SSIS to extract data from HIVE.
  • Designed and developed ETL/ELT jobs across multiple platforms and tools including, Vertica, Hadoop, and Amazon Web Services (AWS).
  • Produced written deliverables for technical designs along with conducting requirements gathering and design sessions with clients.
  • Created Visual Analytic dashboards using Tableau and Adobe Analytics .
  • Used Sqoop to export data into HDFS and Hive.
  • Created Data model to support High performance dashboards and reports using Spotfire.

Environment: T - SQL, PL/SQL, SPSS, SAS, Python, SQL Server (SSIS, SSRS, SSAS), Cognos, Oracle, Teradata, Tableau, PowerBI, OBIEE, UNIX, Spotfire, HDFS, Hadoop, Sqoop, Vertica, Hive.

Confidential, Richardson, TX

Data Modeler/Architect

Responsibilities:
  • Expertise in implementing complex business rules by defining output layouts for Sales, Order Management, Provisioning, Billing, Trouble Ticketing and Financial system reports.
  • Creation of data mart by extracting data from Vertica and Hadoop.
  • Developed an executive dashboard in Tableau with multiple panels and parameters that became a focal point for executives to make decisions.
  • Conducted data preparation, and outlier detection using MS SQL server; built the model using SAS.
  • Developed prediction algorithm using advanced data mining algorithms to classify similar group insurance together to develop sub-markets; each group insurance was divided into subgroups using advanced data mining techniques.
  • Design and develop Mappings to load data from multiple source system databases (DB2, MAINFRAME, Vertica, SQL Server and flat files) in AWS/Azure Environment.
  • Used sqoop to load data into Hadoop, Hive environment.
  • Performed statistical analyses with data sets using statistical packages in SAS and SPSS.
  • Produced written deliverables for technical designs along with conducting requirements gathering and design sessions with clients.
  • Built Dimension Tables and Fact Tables Integrated various sources into the Staging area in Data warehouse.
  • Designed the datawarehouse using Star Schema and created OLAP Cubes for reporting.
  • Led initiative to build statistical models in SAS, Python using historical data to predict
  • Health insurance premiums in several markets. Focused on analyzing the factors affecting Insurance premiums.
  • Created Dashboards using Spotfire, Tableau.
  • Developed prediction algorithm using advanced data mining algorithms to classify similar group insurance together to develop sub-markets; each group insurance was divided into subgroups using advanced data mining techniques.
  • Created advanced reports and dashboards using Spotfire, Cognos, OBIEE and Tableau.
  • Used rest api and javascript api to embed dashboards in applications.
  • Developed Tableau workbooks from multiple data sources using Data Blending.
  • Creating the reports using functionalities like Prompts, Drill Down, Cross Tab and Master Detail for the better understanding of the report, with helps the customer in saving their valuable time.
  • Scheduling of reports according to monthly data load.
  • Performed regression Analysis on the CHURN and customer criteria.
  • Used SAS Enterprise Miner for data mining and e xperience with SAS for statistical modeling .
  • Hands-on modeling and segmentation experience such as CHAID, cluster analysis, decision trees, factor analysis, liner/logistic regression and significance tests.
  • Held role of a Data Scientist and was Responsible for designing and implementing processes and layouts for complex, large- scale data sets used for modeling, data mining, and research purposes.
  • Worked with Data Modeling Tools, Data Mining tools, J2EE and SOA expertise along with UML experience such as writing Use Cases and Activity Diagrams.
  • Proven abilities in creating complex reports by linking data from multiple data providers, using free hand SQL, stored procedures and functionalities like Combined Queries.
  • Experience with Open source Databases such as Hadoop, memchache, MySQL and Mongo.
  • Experience with customer base management, attrition/churn, cost-to-acquire/cost-to-serve, cross-sell/up-sell, Customer lifetime value, retention, loyalty programs, margin per basket/customer, marketing return on investment, micro-targeting/micro-segmentation, segment migration, share of wallet and web analytics.
  • Diagnosed and resolved complex data infrastructure issues including development of methods to identify and resolve data incompatibilities among systems.
  • Performed database indexing, partitioning and data design performance considerations.
  • Used Software as a Service (SaaS) and Performed unit testing, system testing and user acceptances test (UAT) for data model, reports and security management.
  • Involved in gathering, analyzing, and documenting business requirements, functional requirements and data specifications about the financial system for the ETL process and reports.
  • Created ERD, normalized and dimensional models along with conceptual, logical and physical models using Erwin Data Modeler tools, DDL, DML and CA Model Mart.
  • Experience with customer base management, attrition/churn, cost-to-acquire/cost-to-serve, cross-sell/up-sell, Customer lifetime value, retention, loyalty programs, margin per basket/customer, marketing return on investment, micro-targeting/micro-segmentation, segment migration, share of wallet and web analytics.

Environment: DB2, SPSS, SAS, Python, SQL Server (SSIS, SSRS, SSAS), Cognos, Oracle, Teradata, Tableau, PowerBI, OBIEE, Spotfire, UNIX,, HDFS, Hadoop, Sqoop, Vertica, Hive.

Confidential, Plano, TX

Senior ETL/BI Developer

Responsibilities:
  • Developed around 100 Complex Mappings and Reusable Transformations to facilitate daily and Weekly loading of Data.
  • Development of MLOAD scripts.
  • Experience in planning and designing data strategies for very large systems dealing with heterogeneous data sources and unstructured data.
  • Created visualizations to aid in understanding data and built mathematical models using the data.
  • Interpreted and contributed to enterprise data models and data dictionaries.
  • Experience with common social science experimental design, correlations, t-tests and chi squares.
  • Developed various BTEQ scripts to create business logic and process the data.
  • Worked on complete SDLC from Extraction, Transformation and Loading of data.
  • Importing source/target table from respective databases and created reusable transformation, mapplets and mappings using designer and scheduled sessions and batches.
  • Good understanding of Oracle RAC, partitioning, optimization, and tasks typically associated with database conversion along with OLTP databases.
  • Performed ETL with Oracle ODI.
  • Converted packages from DTS to SSIS.
  • Created dashboards and reports using tableau.
  • Defined star schema and staging database enhancements.
  • Experience with common social science experimental design, correlations, t-tests and chi squares.
  • Developed high performance ETL plans to load large volumes of data (60-90+ million records) using Informatica, Ab Initio multifile system and SSIS.
  • Designed developed and tested the DW to extract data from the Administration System. This new system builds to help better design its market goals.
  • Responsible for supporting and troubleshooting Data Quality and Enhancement Issues using quality stage.
  • Proficient in Oracle SQL, PL/SQL & ETL (including complex queries, oracle hints and query optimization techniques).
  • Made use of various transformations in the mapping development. Implemented generic balancing procedure to balance the row counts and hash totals. Unit tested the developed mappings and migrated the mappings across different environments (i.e. Development QA and Production).
  • Used Configuration Files to store the mapping parameters and settings.
  • Involved in all phases of testing mappings and SSIS data flows.
  • Worked on Workflow Manager Tools Task Developer Workflow Designer Worklet Designer. Defined storage parameters partitions and table spaces for Database objects before creating them in the Database. Performed data cleansing and scrubbing in staging area before loading data into the warehouse
  • Used Fast Exports to generate Flat files after the Change data capture has been accomplished which in turn creates a loadable file that is used to load the database.
  • Developed Shell scripts to run the Fast Exports and Bteq Scripts from an HP Unix box.
  • Developed macros and views to integrate complex business rules.
  • Validated the data quality.
  • Extensively worked for the performance of the fast-Export and Bteq utilities.
  • Installed and configured Microsoft IIS Server, Tomcat, Apache, Java Web Server.
  • Source System Analysis and requirements gathering working with different teams to to setup the extraction routines.
  • Moved the mappings from Development environment to Test environment and then to Production Environments.
  • Created multiple unit test cases for the purpose of unit testing the various transformations.
  • Populated error tables as part of the ETL process for the records that failed the migration.
  • Created various transformations including Look Up, Rank, Aggregator, Filter, Router, Joiner, Sequence Generator, Stored Procedure, and Expression.
  • Created batches and session in Informatica and scheduling them to run concurrently or sequentially and provided 24 hour support.
  • Extracting the data from the chapters into flat files using a VB extract utility and FTP them to the national headquaters.
  • Solid SQL experience developing complex queries and Stored Procedures.
  • Used ERWin for Data modeling and reverse engineering.
  • Good experience with Database performance optimization and index design.
  • Creation of mappings from Staging area to clear data holding area with applying various business rules.
  • Creation of Stored Procedures using PL/SQL to capture the error messages and filtered rows, interfaced with system architecture and business customers for requirement gathering.
  • Excellent understanding of OLAP analysis and how OLAP maps to business usage along with Creation of Multidimensional data cubes for OLAP analysis.
  • Worked extensively with Tableau.
  • Created models using Framework Manager.
  • Wrote SAS programs converting questionnaire data to more appropriate formats using SAS Data import, Proc SQL and Data Steps.
  • Performed Statistical quality control of performance measurements.
  • Well versed with STAR Schema.

Environment: SAS, SPSS, Python, R, Informatica, Powerplay, ReportNet, TSQL, SQL Server, FTP, Oracle, Business Objects and Shell Scripts.

Confidential, Dallas, TX

ETL/BI Developer

Responsibilities:
  • Developed ETL mappings, transformations using Informatica Powermart.
  • Created repositories using Informatica Powermart Repository Manager.
  • Worked with sharepoint.
  • Created users and groups and assigned necessary privileges and permission.
  • Implemented source and target definitions in Powermart-Designer
  • Developed mappings using various transformations like update strategy, lookup, stored procedure, Router, Filter, sequence generator, joiner, aggregate transformation and expression.
  • Mappings were designed to cater the slowly changing dimensions.
  • Used various stored procedures in the mappings.
  • Developed several reusable transformations and mapplets that were used in other mappings.
  • Worked on the server manager to run several sessions and batches concurrently and sequentially.
  • Worked with MS Analysis Services and SQL Server Reporting Service.
  • Scheduled the sessions and batches to run at specified intervals.
  • Mappings were designed to the utmost performance by optimizing expressions in expression editor of transformation, filtering source data at the source qualifier, and applying transformation logic.

Environment: Informatica, Python, SPSS, Oracle, SQL Server, PL/SQL, STAR Schema.

We'd love your feedback!