Bi Developer (sql/python/tableau) Resume
Englewood Cliffs, NJ
SUMMARY:
- 6 years of professional experience in Database Management, Statistical modeling, Machine Learning, Data Visualization and Reporting solutions in a variety of industries.
- A profound understanding of Enterprise Data Warehouse (EDW) concepts including OLTP/OLAP, Dimension/Fact table, Star/Snowflake schemas.
- Expert in query relevant data from Microsoft SQL Server, using complexing Structured Query Language and Visual Studio (SSDT), manipulate large unstructured and structured datasets to build insightful solutions to complex problems and present the result in visually engaging and intuitive reports and dashboards.
- Demonstrated experience in SQL writing Queries, Stored Procedures, User - Defined Functions, CTE, Triggers and Transactions.
- Excellent understanding of designing and developing T-SQL queries, ETL packages and business reports using SQL Server Management Studio (SSMS)
- Expertise in Extraction, Transformation and Loading (ETL) data from various sources into Enterprise Data Warehouse and multiple Data Marts using T-SQL, SSIS and Informatica Could on Oracle, Teradata, SQL Server Databases.
- Progressive experience in Tableau Desktop, Tableau Public and Power BI for business analysis, delivering customized ad-hoc reports and performing actionable dashboards for enterprise applications.
- Proficient in utilize parameter, set, calculated field, context filter, cross sheet filter and dashboard action to enhance dashboard performance.
- Advanced skills of SAS programming, familiar with statistic PROC and MACRO
- Expert in Python and its libraries like Pandas, NumPy, Matplotlib, Seaborn, Networks, and Scikit-Learn.
- Strong knowledge in statistical methodologies such as Hypothesis Testing, ANOVA, Experimental Design, Time Series Analysis, Principal Component Analysis (PCA).
- Proficient in Machine Learning algorithms including Linear Regression, Logistic Regression, Elastic Net Regression, K-Means, KNN, Decision Trees, Ridge, Lasso, Random Forests.
- Familiar with Hadoop Ecosystems such as HDFS, MapReduce, Hive and Spark.
- Working experience in Apache Hadoop and Spark frameworks including Hadoop Distributed File System, MapReduce, PySpark and SparkSQL.
- Experienced in generated mathematics and formal reports with a well-designed out layer with Latex.
- Experienced in Cloud Services such as AWS, EMR, Redshift to assist with bid data tools, solve the data storage issue and work on developing solutions.
- Great passion for working with across functional team members and management team at all organizational levels in any emerging business industries such as Healthcare, Commercial and Financial industries and delivering excellent-performed solutions.
TECHNICAL SKILLS:
Programming Languages: Python, T-SQL, SAS, R, SPSS, MySQL, Latex
Databases: MS SQL Server 2016/2014/2012, MS Access, Teradata Database, MongoDB
Tools: SQL Server Management Studio, SQL Server Integration Services (SSIS), SQL Server Data Tools (SSDT), SQL Server Analysis Services (SSAS), AWS, Redshift
Data Visualizations and Reporting: Tableau Desktop/Public, SSRS, Excel, Power BI
Others: MS Project, MS Excel (Pivot Tables/Charts), JMP, SSPS, Gephi
PROFESSIONAL EXPERIENCE:
Confidential, Englewood Cliffs, NJ
BI Developer (SQL/Python/Tableau)
Responsibilities:
- Collected requirements from report developers and designed report demos to satisfying user demand supporting decision making.
- Transferred raw data sources SQL database and other platforms to enrich reporting.
- Collaborated with data engineers and operation team to implement the ETL process, wrote and optimized SQL queries to perform data extraction to fit the analytics requirements.
- Created SQL Store procedure, CTE, Trigger to validate, extract, transform and load data into Data Warehouse and Data Mart.
- Performed data preprocessing and feature engineering for further predictive models including Support Vector Machine, Naive Bayes Classifier to predict potential
- Boosted the performance of regression models by applying polynomial transformation and feature selection.
- Conducted data blending, data preparation using SQL for Tableau consumption and publishing dashboard to Tableau Server.
- Developed Tableau data visualization using Cross tabs, Heat maps, Box and Whisker charts, Scatter Plots, Geographic Maps, Pie Charts, Donut Charts, Bar Charts, and Density Charts
- Utilized quick filters, action filters, parameter controls, achieved functionalities such as dynamic displaying title, commentary, and corresponding records, collapse and expand the table, sub-report linkage, dynamic tooltip displaying, etc.
- Using Tableau Desktop to analyze and obtain insights into large data sets using groups, bins, hierarchies, sorts, sets and filters
- Generated Tableau Public Dashboard with constraints to show specific aspects for a different purpose.
- Involved in automating reports generated a weekly and monthly basis for the client and business sales teams.
Technology Stack: MS SQL server (2012/2014), Excel, T-SQL, SQL Server Data Tools (SSDT), MS office, Tableau Desktop 10.x/2018.x/2019.x, Tableau Server, Python (Numpy, Pandas), Gephi
Confidential, Washington, DC
BI Developer/Data scientist (Python/Spark/Tableau)
Responsibilities:
- Gathered information and requirement from various departments and converted them into documentation for database designing.
- Used Agile methodology and Scrum process for project developing.
- Collected data, drew Entity-Relationship (ER) diagram to identify relationships among all tables, and responsible for creating a database.
- Produced comprehensive analysis report on legacy data, data structure and statistical summary with Python.
- Created database objects like Views, Store Procedure, CTE, Triggers, Functions, etc. using T-SQL on SSMS to store data and maintain database efficiently.
- Involved in the ETL process to ensure that the data warehouse is up to date for reporting purposes and data are well cleaned.
- Used big data tools Spark (Pyspark, SparkSQL) to conduct real-time analysis of the insurance transaction.
- Designed, developed and maintained daily and monthly summary, trending and benchmark reports repository in Tableau Desktop.
- Developed Custom Functions and Complex formulas for reports used calculations, sorts, filters, and sections to enhance the data presented in reports.
- Conducted interactive dashboards with appropriate dashboard level components such as action filters, quick filters, maps, parameters, and used table level calculations to improve visualization of Tableau in terms of performance
- Developed business intelligence dashboards using Tableau producing different summary results based on requirements and role members using donut charts and table contents.
Technology Stack: Microsoft SQL Server 2012, SAS, JMP, Python (PySpark), SSDT, Tableau (Desktop/Server) 9.x/10.x, Visual Studio
Confidential, South Plainfield, NJ
Data Engineer/Tableau Developer (SQL/SAS/Tableau)
Responsibility:
- Worked with project lead and users to understand business processes, gather and document project requirements and translate them into functional and non-functional specifications for BI reports and applications
- Manipulated, analyzed and reported SAS data, transferred SAS data to SQL server.
- Involved in implementation of data viewers, SSIS logging, error configurations of error handling for various events
- Created SSIS packages using execute SQL task, data flow task, data conversion, derived column, lookup, conditional split and aggregate.
- Used SSIS to create ETL packages to Validate, Extract, Transform and Load data into Data Warehouse and Data Marts.
- Improved the performance of T-SQL queries by utilizing index, common table expression (CTE), temp table and removed the unnecessary columns.
- Created Custom SQL queries in Tableau to join tables and views, blending different data connections from different databases.
- Writing custom SQL quires to limit the data sourced into Tableau thus improving Tableau performance.
Technology Stack: MS SQL Server 2012, Visual Studio 2010, SQL Server Integration Services, SAS, MS Excel, Tableau 7.x/8.x
Confidential, Philadelphia, PA
Data Analyst
Responsibilities:
- Analyzed the purpose of stored procedures in the legacy database and converted them into T-SQL.
- Created efficient SQL queries with stored procedure and view to protecting confidential data and optimize the performance of queries using the index and joins.
- Prepared ETL Log Reports and ETL Audit Reports in SQL Server Reporting Services (SSRS) for a project manager to check run status of ETL packages such as ETL name, data flow type, start time, finish time, status and error message
- Identified various data sources, fields and created a data mapping document for the ETL process.
- Imported and exported data from SQL server DB, Excel sheet, flat file by developing SSIS packages and connection manager.
- Designed SSIS packages to handle incremental loads and slowly changing dimensions by lookup and derived column.
- Wrote complex SQL queries using inner join, left join, the case when on tables and views to retrieve data for reporting purpose.
- Developed and validated machine learning models including Ridge and Lasso regression for predicting potential loan default amount.
- Deliver ML algorithms such as Linear Regression, K-means, and KNN to explored and experimented with providing forecasting and analytic insights for various levels of users.
- Performs complex pattern recognition of automotive time series data and forecast demand through the ARMA and ARIMA models and exponential smoothening for multivariate time series data.
- Built classification models using Random Forest, Support Vector Machine and Stacking methods to predict customer attrition.
- Created worksheets and data visualization dashboards with parameters and filters.
- Developed interactive dashboards by using highlights, actions, customized filters to encourage data exploration by users.
- Developed dual-axis line charts with sales and profit gave a certain period.
- Published Tableau reports and dashboards on Tableau Server encouraging business users to explore data.
Technology Stack: MS SQL Server 2014, SQL Server Management Studio, Visual Studio, SQL Server Integration Service, Tableau (Desktop/Server), MS Excel, Python 3.x(Scikit-Learn/Scipy/Numpy/Pandas), Machine Learning (KNN, Clustering, Regressions, Random Forest)