We provide IT Staff Augmentation Services!

Senior Big Data And Bi Lead Resume

4.00/5 (Submit Your Rating)

San Diego, CaliforniA

SUMMARY

  • Senior Big Data Architect Consultant / BI Solutions Architect / Data Management and Cloud Lead
  • 17 years of solid working experience in high - tech industry with a PhD from University of Waterloo.
  • 12 years of experience as a hands-on Solutions Architect/ Team Lead/ Specialist/ Developer in Big Data, Business Intelligence (BI), Artificial Intelligence (AI) and Machine Learning, Cloud Platforms, Data Warehousing, and multi-tiered applications in Scala, Java, Python, PowerShell, and many more.
  • Delivered projects for enterprises such as Starbucks, Confidential, Confidential, the Confidential, the Confidential, Bank of Montreal, General Motors, as well as startups.
  • Expert in Big Data: HDFS, Hive, Spark, Scala, Java, Python, Hadoop, Sqoop, Map/Reduce, Hortonworks, Cloudera, MapR, NoSQL HBase, Cassandra, Kafka, Storm, Spark Streaming, Zeppelin, Kibana, Spark MLib, Mahout, R, NiFi, Falcon, Oozie, Kylin, Atlas, Drill, Solr, ElasticSearch, Ambari,Ranger, Flume, Impala, Pig, HDInsight, SBT, DevOps, Jenkins, Ansible, Git, Maven, Cloudera Navigator, Data Lineage, Kubernetes, Docker.
  • Expert in Machine Learning, Deep Learning, and Data Science, utilizing advanced analytics, artificial intelligence and cognitive services using Spark MLlib Pipelines API, TensorFlow and Keras, Databricks, DSVM, DLVM, R Server, Python, Jupyter notebooks. The solutions include prescriptive (recommendation) and predictive analytics, sentiment analysis, text analytics, vision/image processing, and speech recognition, classification, pattern recognition, anomaly alerting, and regression. Utilized tool sets such as Weka Mahout, Multilayer perceptron classifier (MLPC), the feedforward artificial neural network, scikit-learn, Pandas, Deeplearning4j, H2o, Sparkling Water ML, Caffe2, MxNet etc. Different algorithms K-Means, Random Forest, Gradient Boosting algorithms (GBM, XGBoost and CatBoost) etc.
  • Expert using Azure Cloud (PaaS and IaaS, AaaS), full cycle Cloud Azure experience with full Big Data, Machine Learning Deep Learning, Azure Machine Learning Studio, Azure Power BI, Azure Search, and Elasticsearch, Azure development and deployment ARM templates, Azure CLI, Azure PowerShell, Azure CLI, Automations, Runbooks, Build Tools, CI/CD tools, Azure Data Lake Store, Azure Data Lake Analytics, HDInsight (Managed Hadoop), Databricks, Stream Analytics, Machine Learning Studio, Azure Databricks¸ Azure Data Factory, Azure Data Warehouse, Azure Analysis Services, Event Hubs, Power BI, and Azure Data Catalog.
  • Expert using Google Cloud (GCP), full cycle Cloud experience with full Big Data, Machine Learning Deep Learning, Google Cloud AI, BigQuery, Dataproc, Spark ML, PySpark, Tableau, Google Cloud Search, and Elasticsearch on Google Cloud Platform, Instance templates, Cloud SDK, gcloud CLI, bash, PowerShell, Dataflow, Pipeline, Dataprep, Cloud ML, Automations, Build Tools, CI/CD tools, Stream Analytics and Cloud Pub/Sub, Bigtable, Datastore.
  • Extensive full cycle Cloud AWS experience with full Big Data, Amazon Elastic MapReduce (EMR), Hadoop, Spark, Hive, Pig, Kafka, AWS Management Console, AWS CLI, Amazon EMR File System (EMRFS), collaborative notebooks Apache Zeppelin, Jupyter, deep learning frameworks like Apache MXNet, Elasticsearch and SOLR, Machine Learning and Deep Learning development and deployment. AWS Compute E2C, NoSQL Dynamo DB, FarGate, Lambda, VMware, AWS Developer Tools, AWS Management Tools, Amazon Machine Learning, AWS DeepLens, Amazon Deep Learning AIMs, Amazon TensorFlow on AWS and other components. etc.
  • A full stack solutions architect at enterprise scale. Proficient in end-to-end design spectrum: developing business value justification and strategy, selecting and managing a technology vendor, designing a solution blueprint, envisioning the roadmap, team building, implementation, solution sustainment and DevOps.
  • Defined the data governance strategy, designed cyber security patterns, implemented data standards and procedures across the enterprise; drafted business specific methodology to establish business stakeholder-driven data stewardship through MDM
  • Led EDW projects on Massively Parallel Processing (MPP) Data Warehouse/ Data Lake Appliances such as IBM Netezza, Teradata, Presto, APS (PDW), Oracle Exadata, Customer MDM, Product MDM, DQ Framework.
  • Expert in Enterprise Information Management (EIM) for retail operation. Designed various BI solutions including loyalty programs, card management, POS data management, customer behavioral analysis, store dashboards, finance, ecommerce, and cyber security analytics.
  • Migrated 10s of applications and solutions to the Cloud using Azure, Google Cloud (GCP), Amazon AWS, Talend SnowFlake, Talend Big Data Platform, IBM ICP.
  • Extensively working on SQL Server 7 to 2016, Oracle, SQL, T-SQL, SSIS, SSRS, SSAS, OLTP, OLAP, Multidimensional Cube, MDX, PowerPivot, Tabular Model, SharePoint, PerformancePoint.
  • Demonstrated experience and understanding of the best practices in all aspects of data warehousing (Inmon/Kimball approach). Solid experience in Data Warehouse Development Lifecycle (SDLC) including system analysis, design, implementation, testing, deployment, and maintenance.
  • From Business Case to Data Visualization, I have designed and developed solutions by combining Business Process with Information Technology.
  • Strong knowledge and proven results in Data Warehouse, Data Mart, and Data Lake design including Dimensional Modeling (Star & Snowflake Schemas), ER Modeling, 3 Normal Forms, Normalization and Demoralization, Logical Model and Physical Model, Fact/Dimension/Hierarchy identifications.
  • Years of design and hands on experience in application development using .NET and Service Oriented Architecture (SOA)
  • Authored numerous best practices, technical reports, processes & workflows, and methodology documents to standardize implementation, support infrastructure and sustainment teams, and to promote knowledge sharing. Coached and trained development teams in utilizing the BI framework.

TECHNICAL SKILLS

Big Data: HDFS, Data Lake, Hive, Spark, Scala, Java, Python, Hadoop, Sqoop, Map/Reduce, Hortonworks, Cloudera, MapR, HBase, Cassandra, Kafka, Storm, Spark Streaming, Zeppelin, Kibana, Spark MLib, Mahout, R, NiFi, Falcon, Oozie, Kylin, Atlas, Drill, Solr, ElasticSearch, Ambari, Ranger, Flume, Impala, Pig, HDInsight, SBT, DevOps, Jenkins, Ansible, Git, Maven, Data Virtualization, JBoss, ICP, ELK, Logstash, Kibana, Filebeat, Curator on ICP IBM Private Cloud, Kubernetes, Pods using Helm Charts, Scala SBT, Docker

BI Framework: Strategy and Implementation Plans, Enterprise Metrics, Integration Points, Gap Analysis, BI Portfolio, Performance Management (PM), Analytic and PM Technologies, Defining Business and Decision Process, Building Metadata and Services Centers, Establishing Enterprise Information Management (EIM) Committees, Defining The Role of DW and BI Program Steering Committee, It's Mission, Objectives, Roles and Responsibilities, DAMA DMBOK

Architecture and Data Modeling: Initial Conceptual Solution, Solution Blueprints, Technology Impact Analysis (TIA), Gap Analysis, Technology Roadmap, Dimensional modelling, ER Modelling, Start Schema, Snowflake, Fact, Dimension, Hierarchy, Inmon/ Kimball/ Imhoff, Data Marts, EDW, ERWin 9.5/8.0/7.x, DeZign, Microsoft Visio, Enterprise Architecture (EA), Service Oriented Architecture (SOA), UML, Zachman, TOGAF, Star & Snowflake Schemas, 3 Normal Forms, Normalization and Demoralization, Logical Model and Physical Model, Fact/Dimension/Hierarchy identifications, Data Warehouse Development Lifecycle, Data Mapping, Data Dictionaries

Data Governance: IBM InfoShere MDM, Informatica MDM, MDS, DQS, Profisee Maestro, SAS MDM

Integration and ETL: SSIS/SSRS/SSAS, SQL Server 2014/2012/2008 R2/2008/2005, Informatica PowerCenter, DataStage, Cognos, ETL Mapping design, Data Profiling, Data Validation, Data Migration, Data Cleansing, Data Structure, Data Quality Services (DQS), BIDS, SQL Data Tools (SSDT), Auditing Framework, Execution Plans, ETL Parallel Processing, Error Handling, Custom Scripting, IBM Cognos, InfoSphere DataStage, Informatica PowerCenter, SAP BusinessObjects (BO)

Database: MS-SQL, Oracle, Oracle SQL Developer, TSQL, MDX, DMX, PL/SQL, Stored Procedure, View, Function, Erwin Data Modeler, DB2, PowerDesigner, NoSQL HBase, Cassandra, MongoDB, Access, Excel, FoxPro, Informix, NoSQL, Big-data, Hadoop, Spark, HBase, HDInsight, PDW, PolyBase, Hive, HQL, Map/Reduce, HFS, Alert

Programing Languages: SQL, T-SQL, PL/SQL, C#, WCF, Restful Services, WPF/Silverlight, WF, VB .NET, ASP .NET, ADO.NET, LINQ, MVC, MVVM, MVP, AJAX, HTML 5.0, DHTML, XML, XSL, WSDL, XSD, JSON, Java Script, PowerShell, COM, DCOM, VB Script, UNIX Shell Scripting

Others: Agile, Extreme Programing, RUP, Use Cases, SDLC, TCP/IP, CVS, Microsoft Team Foundation Server (TFS), Github, Jira, bigbucket, Security hardening, Tortoise SVN, SQL*Plus, TOAD, WinSQL, SilverLight, LightSwitch, Kerberos, Single Sign-On, Datazen, One-Key.

Integration: Windows API, Biztalk, SOA, WCF, SSIS

Data Access: ADO.NET, LINQ, Entity Framework, Microsoft Enterprise Library, OLE DB, Oracle Data Provider, MS OLAP, SQL Master Data Services (MDS), StreamInsight

Methodologies: Test driven programming, Agile software development, Extreme Programming (XP) Microsoft .NET Framework (from 1.0 to 4.0), C#, Visual Basic .NET, VB .NET, ADO .NET, WinFX including Windows communication foundation (WCF), windows workflow (WF), windows presentation foundation (WPF), XAML, XML, HTML, HTML5, Java J2EE, Spring Framework, Scala, JavaScript, AJAX, RESTful services, Payment Card Industry (PCI), Image Processing Visual C++ (MFC, SDKs, COM, DCOM, ATL ActiveXs), VB, C++, Perl, VHDL, Verilog, Shell, Skill, Ocean, SystemC

Version control tools: Source Safe, Team Foundation Version Control, (TFVC), Subversion Tortoise SVN

Cloud Computing: Azure Cloud, Cloud Formation, ARM templates, AWS CLI, Azure PowerShell, Azure CLI, Automations, Runbooks, Build Tools, CI/CD tools (Chef, Jenkins, Ansible, and Kubernetes, Docker, container orchestration), DevOps and Agile principals, Azure Data Lake Store (ADLS), Azure Data Lake Analytics (ADLA), HDInsight (Managed Hadoop), Machine Learning Studio (ML), Azure Databricks, Azure Data Factory (ADF), Azure Data Warehouse (ADW), Azure Analysis Services (AAS), Power BI, and Azure Data Catalog, Amazon EC2, NoSQL DynamoDB

SharePoint: SharePoint 2013/2010/2007/2003 , Multi-machine SharePoint Farm Architecture, Setup, Configuration, Load Balancing, Clustering, Backup Plans, Web Part and module development, Collaboration, Social, Search, Web Content Management, Enterprise Content Management, App Management, PerformancePoint and PowerPivot, PowerView, Application Federation, Secure Store Application, Business Connectivity, Usage Reports, SharePoint Designer, Dashboard Designer, PerformancePoint, dashboard, charts, KPI, Scorecards, reports, filters, Excel Services, PowerPivot Services, Web Analytics, Static Analysis, Hit Counters, Custom Development, PowerShell, SharePoint API, Object Model, web parts web services, workflows, Content Management, site collections/structure

Internet Development: ASP .NET, MVC, Sliverlight, HTML, DHTML, Web services for marketing and financial applications, AJAX, ASP, JavaScript and VB Script, XML, Microsoft Internet Information Server (IIS), Microsoft E-Commerce, PHP, Webload

Mobile Development: Android, iOS, Windows

Operating System Used: Windows, UNIX, Windows Azure, Linux, Android, iOS, Windows Mobile, MS-DOS

Hardware and Simulation: Matlab, Cadence Spectre, Spice, Eldo, ANSYS

Algorithms: Genetic algorithm, simulated annealing based algorithms, heuristic search, binary search, quick sort

Automation and Scripting: VB, Perl, Unix Shell

IDE: Eclipse, IntelliJ, Visual Studio

PROFESSIONAL EXPERIENCE

Senior Big Data and BI Lead 

Confidential, San Diego, California

Responsibilities:

  • Designed and developed data engineering solutions for a top Healthcare / Pharmaceutical company with $100 billion asset, managed on a secure and scalable analytics platform.
  • Led projects to ingest, enrich and curate data to help the company be compliant with GxP guidelines such as Good Clinical Practices (GCP), Good Lab Practice (GLP), Good Manufacturing Practice (GMP), and FDA Regulatory Expectation. The data solutions facilitates a better quality monitoring, primary record management, validation as well as product, incident, test, change, configuration, and document management.
  • Designed and developed solutions by utilizing a set of big data tools such as Spark clusters, Hadoop, YARN, Hortonworks, Hive, Sqoop, NoSQL HBase, oozie, Spark Stream, Kafka, Atlas, Ranger, data catalog, Python, Scala, Jupyter notebooks, and Data visualization.
  • Migrated on-premises clusters to Microsoft Azure Cloud and enabled data scientists perform machine learning and advanced analytics by utilizing Azure Data Factory Pipelines, Data Lakes, Blobs, Catalogs, Keyvaults, HDInsight, Databricks, Azure ML Studio, PowerShell, Automations, Runbooks, CI/CD tools (Chef, Jenkins, Ansible, and Kubernetes, Docker, container orchestration), DevOps.

Senior Big Data and BI Lead / Solutions Architect - Consultant

Confidential, Seattle, Washington

Responsibilities:

  • Led BI projects utilizing big Data, machine learning, and cloud platforms to optimize inventory, customer experience, predictive maintenance, and supply chain for 28,000 locations spread across 77 countries.
  • Designed and developed a big data platform and data pipelines on large clusters utilizing Spark, Scala, SBT, Hadoop, YARN, MapR, Hive, Flume, Sqoop, Map/Reduce, Pig, NoSQL HBase, oozie, Spark Stream, Kafka and Data visualization. The project enabled continuous data processing for data exploration, data preparation and real-time business analytics in accordance with Starbucks guiding principles.
  • Designed and developed solutions for exploitation of batch and real-time machine learning and deep Learning frameworks algorithms for advanced predictive analytics driven reporting and anomaly alerting. The solutions utilize tools such as Spark MLlib Pipelines API, TensorFlow and Keras, Databricks, DSVM, DLVM, R Server, Python, Jupyter notebooks, Spark, HDFS, Scala, SBT. Some of the developments were based on Multilayer perceptron classifier (MLPC) which is a classifier based on the feedforward artificial neural network. Also created other prototypes using Python (scikit-learn, scipy, pandas, numpy), R, PySpark and Scala libraries like scikit-learn, Pandas, Deeplearning4j, Sparkling Water ML, Caffe2, MxNet etc. Different algorithms were used like K-Means, Random Forest, Gradient Boosting algorithms (GBM, XGBoost, XGBoost and CatBoost). The solutions applied advanced machine learning and AI technologies for inventory and supply chain management in retail domains.
  • Led cloud compute platforms for enabling data scientists by utilizing Microsoft Azure, Cloud Formation, ARM templates, Azure CLI, Azure PowerShell, Azure CLI, Automations, Runbooks, Build Tools, CI/CD tools (Chef, Jenkins, Ansible, and Kubernetes, Docker, container orchestration), DevOps and Agile principals. Also, developed automated solutions, as one-click workflows, for automatically provisioning and configuration of the end to end analytics platform with all the associated resources such as HDInsight, Databricks, CosmosDB, Azure Data Warehouse, Polybase, AWS, SQL Server, ADLS, blob, Functions, Automation Runbooks, Data Factory pipelines, Virtual Networks, Active Directory, KeyVaults, and Resource Groups. In addition, designed and developed security monitoring and governance solutions that manages permissions in Azure Data Lake Stores and manages the costs and policies across the entire subscription.

Senior BI Lead / Solutions Architect

Confidential, San Diego, California

Responsibilities:

  • Led Big Data and BI projects to analyze data for a $100 Billion investment, managed by Confidential (Ontario Municipal Employees Retirement System). The projects ingest, transform, and analyze data for Capital Market, Private Equity Market, and Pension investment management. Created a big data framework architecture and implemented solutions using Microsoft Azure, Cloud Formation, ARM templates, AWS CLI, Azure PowerShell, Azure CLI, Automations, Runbooks, Build Tools, CI/CD tools, Azure Data Lake Store, Azure Data Lake Analytics, HDInsight (Managed Hadoop), Machine Learning Studio, Azure Databricks, Azure Data Factory, Azure Data Warehouse, Polybase, Azure Analysis Services, Azure Data Catalog, Talend SnowFlake, Talend Big Data Platform, and Power BI.
  • I also performed stream analytics using machine learning anomaly detection for cyber security outages using Google Cloud (GCP), BigQuery, Dataproc, Spark ML, PySpark, Tableau, Instance templates, Cloud SDK, gcloud CLI, bash, Dataflow, Pipeline, Dataprep, Cloud ML, Automations, Build Tools, CI/CD tools, Stream Analytics and Cloud Pub/Sub, Bigtable, Datastore, Machine Learning, TensorFlow.
  • Designed and developed a Big Data project utilizing Spark, Kafka, Java, MapR, Hadoop, Hive, Impala, Flume, Sqoop, Map/Reduce, and Pig. The project collects information from capital market, performs advanced analytics for risk management. Subsequently, the unstructured data is processed to feed the BI layer.
  • Designed solution, performed data modeling, and automated the data flow for Cyber Security and Incident Management System (ServiceNow) by running U-SQL scripts on Azure Data Lake and scheduling the data pipelines using data factory. Provisioned Azure environments by using Azure PowerShell and Resource Manager to manage the resources.
  • Successfully installed and deployed an entire IBM Cloud Private ICP Cluster then implemented and deployed ELK Elasticsearch, Logstash, Kibana, Filebeat, Kafka, Zookeeper, NoSQL Cassandra, Curator on ICP IBM Private Cloud, Kubernetes, Pods using Helm Charts, Scala SBT. IBM Cloud and IBM Cloud Private (ICP) is a Distributed Container based Architecture, Docker, Docker CLI, Kubernetes, Kubernetes CLI, Pods, Pods deployments. Service Deployments, Ingress, Helm Charts, Helm Charts CLI.

Senior Big Data / BI Solutions Architect - Consultant

Confidential

Responsibilities:

  • Designed and developed a Big Data project on Teradata utilizing Spark, Kafka, Java, Hortonworks, Hadoop, Hive, Impala, Flume, Sqoop, Map/Reduce, and Pig. The project collects information from cameras and sensors installed on the highways, performs advanced analytics to capture the information.
  • Led BI/DW projects for Ontario's driver examination services. The solution consolidates the information from 55 Drive Test centers as well as 39 travel point locations. Designed the full stack BI solution including ETL, Staging, DW, Dashboard, and reporting layer in Power BI.
  • Designed and developed an advanced analytical solutions that utilizes Machine Learning and Deep Learning frameworks for Ontario's driver examination services. The solutions utilize tools such as Machine Learning Studio, DLVM, R Server, Python, Jupyter notebooks, Azure Data Lake, Azure SQL database, Azure SQL Data Warehouse, Polybase, and many other data science and ML tools, Spark MLlib Pipelines API, TensorFlow and Keras, Databricks, DSVM, DLVM, R Server, Python, Jupyter notebooks.
  • Analyzed and prepared various datasets, ingested them into Google Cloud (GCP) utilizing Google Cloud AI, BigQuery, Dataproc, Spark ML, PySpark, Dataflow, Pipeline, Dataprep, Cloud ML, Automations, Bigtable, built data model, and applied various Machine Learning algorithms to find the most important parameters impacting the wait time at the centers and to predict the wait time at the centers. These algorithms include regression, Two-class classification, Multi-class classification, and anomaly detection. I also performed What-If analysis, proposed actions to be taken to reduce the wait time.
  • Designed and developed Artificial Intelligence (AI) applications with Natural Language Processing (NLP) - With Cognitive Services API's. The solution include text analytics, Vision, and Speech recognition. The project collects information from cameras and sensors installed on the highways, performs advanced analytics to capture the information. Subsequently, the unstructured data is processed to feed the BI layer.
  • Designed and developed fast scalable search system using a multicloud solution with Azure and AWS utilizing ElasticSearch for incident and ticket management. The high speed search gets its data from a Machine Learning system with Key Phrase extraction capabilities. Subsequently, user can search and perform text mining through unstructured text in a large scale data source.
  • Established a roadmap for migrating legacy information warehouse to a modern high performance data warehouse running on multiple DW appliances. Drafted a BI/DW prioritized implementation roadmap while taking input from organization, business, and IT strategy documentation
  • Created high-level architecture, solution blueprints, technology infrastructure implementation plan, integration plan, and disaster recovery plan. Created architecture artifacts and provided subject matter expertise in support of a technology selection process.
  • Delivered project utilizing full cycle Cloud AWS experience with full Big Data, Amazon Elastic MapReduce (EMR), Hadoop, Spark, Hive, Pig, Kafka, AWS Management Console, AWS CLI, Amazon EMR File System (EMRFS), collaborative notebooks Apache Zeppelin, Jupyter, deep learning frameworks like Apache MXNet, Elasticsearch and SOLR, Machine Learning and Deep Learning development and deployment. AWS Compute E2C, FarGate, Lambda, VMware, AWS Developer Tools, AWS Management Tools, Amazon Machine Learning, AWS DeepLens, Amazon Deep Learning AIMs, Amazon TensorFlow on AWS and other components. etc.

Senior BI Lead/ Solution Architect Consultant

Confidential

Responsibilities:

  • Led a big data project on large clusters utilizing Spark, Scala, Hadoop, Hortonworks, Cloudera, Hive, Impala, Flume, Sqoop, Map/Reduce, Pig, HDInsight, NoSQL HBase, oozie, Data visualization, and facilitating real-time data analysis by data scientist.
  • Designed and developed a sales prediction engine for 500 stores to forecast the sales across the country using Amazon AWS services including EC2, S3, DynamoDB, Presto, Redshift, EMR, Data Pipeline, Kinesis Analytics, Elasticsearch Service.
  • Led EDW projects on Massively Parallel Processing (MPP) Data Warehouse Appliances such as IBM Netezza, Teradata, APS (PDW), Oracle Exadata, Java
  • Defined the data governance strategy, designed security patterns, implemented data standards and procedures across the enterprise; drafted business specific methodology to establish business stakeholder-driven data stewardship through MDM
  • Designed Enterprise Information Management (EIM) solutions for retail operation. Led technical teams, provided estimation, planed resources, and designed various BI solutions including loyalty programs, card management, POS data management, customer behavioral analysis, store dashboards, finance, ecommerce, cyber security analytic.
  • Drafted a BI/DW prioritized implementation roadmap while taking input from internal divisional service plans, business and IT strategy documentation, as well as corporate BI Strategy and the Financial Planning and Reporting System
  • Created high-level architecture, solution blueprints, technology infrastructure implementation plan, integration plan, and disaster recovery plan. Created architecture artifacts and provided subject matter expertise in support of a technology selection process
  • Designed a Scorecard and Strategy Management BI solution with capabilities aimed at communicating strategic goals across the organization and monitoring progress over time. Scorecard and Strategy Management provided capabilities to establish specific goals, defined how to measure success, and communicated information down the entire organization to increase organizational effectiveness and productivity through improved transparency, processes and technology
  • Conducted and documented business stakeholder interviews and workshops
  • Conducted BI maturity assessment of the organization. Architected DW&BI Program Structure, defined the role of DW&BI Program Steering Committee, it's mission, objectives, roles and responsibilities, monitored regular improvements to help manage risks, evaluate trends, and develop capacity and capability to achieve the Program mission
  • Led a team of 16 individuals to plan, design, and migrate a portfolio of 6 projects to Azure Cloud. Was involved in the full cycle of vendor selection, evaluated RFP responses, requirement gathering, design, and development of this project. The migration included different aspects of the project from front, backend, and integration. According to the requirement, we selected Azure cloud for this project and utilized cutting edge utilities to perform the migration.

Senior BI Lead/ Solution Architect Consultant

Confidential

Responsibilities:

  • Led a big data project to be hosted on Azure cloud. Led the design and development of big data solution to conduct sentiment analysis for all schools across Ontario. The solution merges feeds from social media with school board structured data. The data then is analyzed by multiple dimensions including geographical, parent income, and crime rate. The analysis provides valuable insight and is used in planning, early detection of issues, and enhancing school performance. The tools utilized include Hadoop, HDInsight, PDW, Map/Reduce jobs, Java, Hive, and Sqoop.
  • Led the design and development of BI solution, ETL processes, and data mapping using SQL server, Master Data Services (MDS), Data Quality Services (DQS), SSIS to extract data from ITIL and Process Implementation tools such as Remedy and HP OVSD. The process loads the asset management information including server hardware, software, and application data and transform and load the data into data warehouse. Subsequently, designed and developed reports that are accessible through a SharePoint portal. Developed several BI components for implementation of new policies, process improvements and detailed work instructions that integrates application management processes with existing ITIL based processes such as incident management, change, knowledge, IT financial, and configuration management.
  • Led the design and development of a SQL Server SSAS Analysis cube utilizing star schema with complex MDX calculated measures, named sets and KPIs, KPPIs to present an analytical view for the data and data quality with multiple dimensions.
  • Led the design and development of data visualization using Microsoft Power View and Power Map to represent the server information across multiple data centers in Canada.
  • Was involved in the full cycle of vendor selection, evaluated RFP responses, requirement gathering, design, and development of the projects.

Senior BI Solutions Architect/ Lead/ Senior BI Specialist

Confidential

Responsibilities:

  • Led the design and development of 311 data warehouse and business intelligence solution that provide a consolidated view for the data for all services that the City provides to the citizens. The solution extracts data from Lagan ECM, different divisions including 311 Toronto, Toronto Water, Solid Waste Management, Municipal Licensing & Standards, Forestry & Recreation, Transportation Services and Parks. The extracted data, then, are corrected and cleansed according to business rules and presented as a single analytical view to the users such as Councillors, City senior management, and operational managers.
  • Led the design and development of a big data solutions by utilizing Hadoop/HDInsight, PDW, Map/Reduce jobs, Hive, and Sqoop. The solutions measure the performance of 311 call centre by analyzing the data gathered from Cisco/IPCC telephone systems. Integrated with Cisco Verint for VOIP call recording, quality monitoring (QM), and speech analytics. The BI solution improves the communications between the City and the customers. The real-time data visualization shows a snapshot of the interactions between the CSRs and the customers at any given time.
  • Led the design and development of the Workforce Management (WFM) data warehouse and BI solution to optimize adherence and attendance in the contact centre. The predictive analytic component accurately forecasts the number of CSRs needed in the call centre to fulfill the services.
  • Led the design and development of an efficient BI auditing framework that collects the data from packages being executed and used in data flows, row counters, versioning, and error handling. The framework is crucial for monitoring, timing, troubleshooting, and auditing. Also, developed Stored Procedures, Views, and Functions for the framework to automate logging the information and error handling in the packages.
  • Led the design and development of ETL processes and data mapping using SQL server, Master Data Services (MDS), SSIS to extract data from Lagan ECM and division data sources including SQL server and oracle databases, flat files, and excel sheets. The data, then, is transformed and loaded into a data warehouse for reporting.
  • Led the design and development of data quality ETL packages to correct and cleanse the data and enhance the quality of consolidated data. Wrote hundreds lines of .NET C# code, embedded in the packages, to create a rules engine that loads business rules and apply them to the data efficiently. In addition, the data quality issues are mapped for reporting purposes.
  • Led the design and development of a SQL Server SSAS Analysis cube utilizing star schema with complex MDX calculated measures, named sets and KPIs to present an analytical view for the data and data quality with multiple dimensions.
  • Led the design and development of map application and report using ASP.NET/C# web application. The application loads the data from the data warehouse, combines it with geographical information, and displays the data on a map. The application communicates through restful mapping services and uses client side scripts (JavaScript and AJAX) to improve performance and user experience.
  • Led the design and development of SharePoint web application, site collection, dashboard, and libraries. To present the analytical reports in a user-friendly and interactive way, SharePoint and dashboard designer 2010 are used to create a site that hosts PerformancePoint charts, filters, and integrated SSRS reports. The dashboard is connected to an SSAS cube and dynamically applies the filters to the data shown on the charts.
  • Prepared system architecture documents and designed for high-availability clusters. Also, he configured multi-machine SharePoint farm through windows clusters to ensure reports and dashboards are always available in case of a failure. Also, configured distributed reporting services to be integrated with the farm.
  • Led design of a Java application using Spring Framework to communicate, create and update the booking for the City facilities.
  • Led the configuration of access levels for different groups and created security roles within SharePoint.
  • Led the enhancement of the security of map application through SSL, security certificates, and windows authentication in IIS.
  • Authored naming conventions, BI best practices, and development and deployment guidelines. The guidelines also provide steps for SharePoint installation, setup, backup, and recovering the system to help sustainment team promote the solutions to QA and production.

We'd love your feedback!