Developer /analyst Resume
Toronto, ON
SUMMARY:
- As a Developer/Analyst with over 10 years of experience in designing, developing and delivering enterprise applications on the Microsoft Data Platform, I am recognized by managers, and peers as a dedicated and reliable performer. I am capable of demonstrating innovative ideas and effectively communicate with team members to ensure quality and timely project completion. My experience spans Web Development, Database Development (OLTP and OLAP), Enterprise Data Warehousing, ETL, Data Visualization, Reporting and Data Mining .
- Experience with working on the Microsoft Azure Cloud platform for development and deployment.
- Performed migration of data from on premise SQL Servers to Azure SQL Database using SSMS, SSIS, Replication and Azure Data Factory .
- Experienced creating Azure HDInsight Hadoop clusters with Azure PowerShell to process Big Data workloads.
- Created Azure Data Factory pipelines, transformations and schedules to move data from on premise sources to Azure Blob Storage, HDFS and Azure SQL Database .
- Experienced working with various RDBMS like SQL Server, Oracle, DB2 and Apache HBase NoSQL database.
- Experienced working with both OLTP and OLAP database systems using Microsoft SQL Server .
- Experienced working in Enterprise Data Warehouse projects using Star Schema Dimensional Models .
- Experienced building ETL packages for various Data Marts and Operational Data Stores using SSIS .
- Experienced working with Data Quality Services and Master Data Services in an Enterprise Data Warehouse environment.
- Proficient in creating Tabular and Multi - Dimensional data models using Microsoft SQL Server Analysis Services .
- Experienced in MDX and DAX in performing queries and calculations against SSAS Multi-Dimensional and Tabular Data Models .
- Proficient at working with Hive Data Definition Language, Hive Database, Hive Tables, Hive Partitions, Hive Buckets, Hive Views, Data Manipulations, Aggregation, User Defined Functions etc. on Apache Hive .
- Created Apache Spark Clusters on Azure HDInsight to work with Big Data workloads and analyzed data using the Spark’s API for Scala and Python as well as Spark SQL .
- Experienced designing publishing sites using SharePoint 2010 and Office 365 .
- Experienced developing n-tiered Web Applications using ASP.NET Webforms and MVC .
- Experienced being part of an Agile development team.
- Proficient in performing MapReduce operations and various other Transformations and Actions on Spark RDDs and Dataframes using PySpark and Scala .
- Proficient in developing Machine Learning Models using Scikit-Learn, Azure Machine Learning, Apache Spark and H2O Sparkling water.
- Proficient in performing Exploratory Data Analysis using Python and R as well as plotting various data visualizations using Matplotlib, Seaborn and ggplot2 .
- Early adopter and avid follower of Microsoft Power BI for creating visuals and reports with Data sourced from SSAS Multidimensional and Tabular Data models.
PROFESSIONAL EXPERIENCE:
Confidential, Toronto, ON
Developer /Analyst
Responsibilities:
- As a member of an Agile team attended Sprint Planning, Daily scrum meetings, Sprint Reviews and Sprint Retrospectives .
- Streamlined Sprint tasks to optimize time and resource to facilitate completion of task as per schedule.
- Kept team members and business stakeholder up to date regarding work progress.
- Took on a fresh challenge to perform R&D for an Enterprise Cloud Strategy to adopt Microsoft Cloud as an alternative platform for Enterprise application development and deployments.
- Created Proof of Concept applications on platforms like the Azure SQL Database, Azure Web Apps and Azure Machine Learning .
- Migrated on premise SQL Server Databases to Azure SQL Databases.
- Several internal infrastructure units were decommissioned after migrating applications and databases to Microsoft Azure .
- Created Machine Learning Models for various tasks like Classification, Regression on the Azure Machine Learning platform and created client applications to utilize the predictive models created using Azure Machine Learning .
- Performed ETL, Data Preparation, Exploratory Data Analysis, Feature Extraction, Feature Engineering, Dimensionality Reduction, Modeling, Training, Hyper Parameter Tuning, K-Fold Cross Validation and deployment of Machine Learning models.
- Performed feature selection using techniques like Filter Based Feature Selection, Fischer Linear Discriminant Analysis, Sequential Backward Selection, Permutation based Feature Selection, Logistic Regression with L1 regularization, Random Forest Classifier etc .
- Used Python extensively to clean and transform data and to perform statistical analysis like Summary Statistics, Correlation Analysis .
- Used Matplotlib to create visualizations like Barplots, Histograms, Scatterplots, Boxplots, 3-Dimensional plots, Learning Curves, Validation Curves.
- Used Seaborn to create visualizations like Histograms, KDE, Pairplots, Pairgrid, Faceted Histograms, Factorplots, Facegrid etc.
- Performed data compression via Dimensionality Reduction techniques like PCA, Kernel PCA and LDA .
- Created model Pipelines, evaluated models with k-fold cross validation and tuned hyperparameters using grid search with Scikit-Learn .
- Created ensemble models using majority voting, bagging and Adaboosting techniques .
- Created Flask web application to operationalize the Machine learning model created in Scikit-learn.
- Created Azure Data Factory pipelines to move data between Azure Blob Storage and various other Data Sources.
- Created Apache Spark Clusters on Azure HDInsight clusters with Windows Azure Blob Storage WASB as the primary file system HDFS .
- Used Azure SQL Database as Apache Hive Metastore for reusability of Hive artifacts on the Hadoop Cluster.
- Wrote Hive Queries to analyze and perform ETL on the Data on Apache Hive .
- Performed numerous Spark transforms and actions during data preparation.
- Created Spark MLlib Pipelines consisting of Estimators, Transformers and Parameters .
- Utilized Feature Extraction algorithms like TF-IDF, Word2Vec, CountVectorizer, VectorSlicer from the MLlib library.
- Used Algorithms like Tokenizer, StopwordsRemover, PCA, OneHotEncoder, VectorIndexer, Normalizer, StandardScaler, MinMaxScaler, Bucketizer etc to transform features.
- Created SSIS Packages to load Fact and Dimension tables in the Enterprise Data Warehouse, to load Operational Data Stores for ad-hoc reporting and to send and receive data feeds from various vendors and business partners,
- Create Multidimensional and Tabular Data Models on SQL Server Analysis Services using SQL Server Data Tools.
- Used MDX and DAX to create custom Measures and Calculations against the Multidimensional and Tabular Data Models .
- Used Power BI to create various visualizations like Bar Chart, Column Chart, Line Chart, Pie Chart, Scatter Charts etc.
- Used Power BI to create slicers and various filters at Report, Page and Visualization levels.
- Used Power BI Service to build dashboards and share reports with users.
- Used Enterprise Data Gateway to access data from on premise Data Stores and to display it on the Power BI Service .
Environment: Agile, Version One, Visual Studio Online, Python, Pandas, Numpy, Scipy, Seaborn, Matplotlib, Scikit-Learn, Hadoop, Apache Hive, Apache Spark, H2O, HDInsight, Azure, Azure Machine Learning, Scala, Jupyter Notebooks, Spyder, Eclipse, Intellij IDEA, C#, Power BI, Excel, Tableau, Office 365, Active Directory, PowerShell, SQL Server, Azure SQL Database, Azure SQL Data Warehouse, SSAS, SSIS, MDX, DAX, Oracle, DB2,Visual Studio, TFS, VSTS, Windows Server 2008,Windows Server 2012.
Confidential, Toledo, OH
Senior Consultant
Responsibilities:
- Meeting with the business users to gather software requirements and worked closely with the Architecture team to design Data Models .
- Created Logical and Physical Data Models for OLAP and OLTP systems.
- Normalizing the tables/relationships to arrive at effective Relational Schemas.
- Conducted meetings with various stakeholders to discuss security, compliance and usability factors of the applications.
- Implementation of Business Rules in the database using references, constraints & triggers.
- Developed DDL scripts to create the database schema and database objects.
- Performed Dimensional Modeling of Datamart using star and snowflake schemas.
- Identifying Dimensions and Facts and the required grain for the Fact tables.
- Design of Operational Data Stores for real-time reporting.
- Perform analysis of business requirements and transformed them into technical requirements for ETL and Reporting needs.
- Building Extract, Transform and Load (ETL) solutions for data warehousing using SQL Server Integration Service (SSIS).
- Used transformations like SCD, CDC, Data Conversion, Conditional Split, Merge Join, Derived Column, Lookup, Cache Transform and Union All etc., to convert raw data into required data format meeting the business requirements.
- Extensively worked on Control Flow Tasks such as For each Loop Container, For Loop Container, Data Flow Task, File System Task, Script Task, Execute SQL Task, FTP Task, Execute Process Task, Send Mail Task etc.
- Efficiently used Data Sources and Destinations such as Excel Source, Flat File Source, OLE DB Source, Excel Destination, Flat File Destination, OLE DB Destination, Raw File Destination, SQL Server Destination.
- Designing and implementing robust Exception Handling and Logging in the packages.
- Designed OLAP Cubes, Data Source Views (DSV) using SSAS with the data from the Enterprise Data Mart.
- Created numerous calculations on the Cube using MDX for use in reports.
- Developing efficient and sometimes complex Stored Procedures, Functions, Triggers, Views using T-SQL.
- Used Redgate tools to compare database schemas, data and to generate data for load testing the Database.
- Optimizing query performance by making modifications in T-SQL adding or removing indexes where it made sense by taking hints from the Query Plan.
- Report Development with SSRS (tabular reports, parameterized reports, drilldown, drill through, sub reports, charts etc.)
- Deploying reports to the Report Server, creating report schedules and subscriptions, managing and securing report definitions and data sources.
- Worked in integrating SQL Server BI with SharePoint BI and using SSSRS in SharePoint mode.
- Creating Performance Point Dashboards on SharePoint using Dashboard Designer.
- Supporting QA testing, involved in troubleshooting, debugging and fixing the codes/scripts.
Environment: SQL Server, Oracle, Windows 2008 R2, Visual Studio, Team Foundation Server, Crystal Reports, SSRS, SSIS, SSAS, MDX, Redgate, C#, SQL, SharePoint 2010, Sharepoint Web parts, Sharepoint Workflows, Excel Service, Performance Point Service, MS Visio.
Confidential, Jersey City, NJ
Consultant Responsibilities:
- Extensively worked with ASP.NET 2.0, ADO.NET 2.0, Visual C#, and SQL Server 2000/2005 .
- Used ASP.NET Controls, Master Pages, CSS and HTML for web development projects.
- Used ADO.NET 2.0 components like DataSets, Data Table, TableAdapter, SqlDataProvider for read/ write data with the SQL Server Back End.
- Used C# language for coding in the Business Logic Layer of the 3-tier application architecture.
- Used Windows Authentication in intranet sites and queried Active Directory to get user data.
- Used various Validation Controls and JavaScript for validations.
- Used Log4Net to log application errors for support purposes.
- Wrote Stored Procedures, Functions and Triggers for using T-SQL .
- Created data objects in the business layer for retrieving data from database through the DAL.
- Created DTS packages to perform data loads and extracts from different source systems and file formats.
- Created reports like Drill-Down, Drill-Through, Chart, Sub Reports using SQL Server Reporting Services (SSRS).
- Supported the QA testing and end user training programs with inputs to creating User Help files.
Environment: C#, ASP.NET, AJAX, ADO.NET, XML, WSDL, SOAP, Web Services, SQL Server 2005, Oracle, DB2, BIDS, SSRS, IIS 6.0, CSS, HTML, JavaScript, Visual Studio 2005, Windows 2003, Visual Source Safe.