We provide IT Staff Augmentation Services!

Data Analyst Resume

5.00/5 (Submit Your Rating)

Newark, NJ

PROFESSIONAL SUMMARY:

  • Above 8+ years of experience in Data Analysis, Machine Learning, Data mining with large data sets of Structured and Unstructured data, Data Acquisition, Data Validation, Predictive modeling, Data Visualization.
  • Experienced in lasted BI tools like Tableau, Power BI, Qlik Sense, Qlik View.
  • Expertise in transforming business requirements into building models, designing algorithms, developingdatamining and reporting solutions that scales across massive volume of unstructureddataand structured.
  • Proficient in Machine Learning techniques (Decision Trees, Linear, Logistics, Random Forest, SVM, Bayesian, XG Boost, K - Nearest Neighbors) and Statistical Modeling in Forecasting/ Predictive Analytics, Segmentation methodologies, Regression based models, Hypothesis testing, Factor analysis/ PCA, Ensembles.
  • Experience in designing visualizations using Tableau software and Storyline on web and desktop platforms, publishing and presenting dashboards.
  • Strong experience in Software Development Life Cycle (SDLC) including Requirements Analysis, Design Specification and Testing as per Cycle in both Waterfall and Agile methodologies.
  • Proficient in Data visualization tools such as Tableau, Plotly,PythonMatplotlib and Seaborn.
  • Familiar with Hadoop Ecosystem such as HDFS, HBase, Hive, Pig and Oozie.
  • Experienced in building models by using Spark (PySpark, SparkSQL, Spark MLLib, and Spark ML).
  • Experienced in Cloud Services such as AWS EC2, EMR, RDS, S3 to assist with big data tools, solve the data storage issue and work on deployment solution.
  • Data scientist and mentor the team to prepare new POCs and models for healthcare customers.
  • Worked and extracteddatafrom various database sources like Oracle, SQL Server and Teradata.
  • Experience in foundational machine learning models and concepts( Regression, boosting, GBM, NNs, HMMs, CRFs, MRFs).
  • Skilled in System Analysis, DimensionalDataModeling, Database Design and implementing RDBMS specific features.
  • Developed a generic model for predicting repayment of debt owed in the healthcare, large commercial, and government sectors.
  • Facilitated and helped translate complex quantitative methods into simplified solutions for users.
  • Knowledge of working with Proof of Concepts and gap analysis and gathered necessarydatafor analysis from different sources, prepareddatafordataexploration usingdatamunging.
  • Git,Java,MySQL,MongoDB,Neo4J,AngularJS,SPSS,Tableau.
  • Excellent knowledge of Hadoop Ecosystem and Big Data tools as Pig, Hive &Spark.
  • Worked on different data formats such as JSON, XML and performed machinelearningalgorithms in Python.

TECHNICAL SKILLS:

Exploratory Data Analysis: Univariate/Multivariate Outlier detection, Missing value imputation, Histograms/Density estimation, EDA in Tableau

Supervised Learning: Linear/Logistic Regression, Lasso, Ridge, Elastic Nets, Decision Trees, Ensemble Methods, Random Forests, Support Vector Machines, Gradient Boosting, XGB, Deep Neural Networks, Bayesian Learning

Unsupervised Learning: Principal Component Analysis, Association Rules, Factor Analysis, K-Means, Hierarchical Clustering, Gaussian Mixture Models, Market Basket Analysis, Collaborative Filtering and Low Rank Matrix Factorization

Sampling Methods: Bootstrap sampling methods and Stratified sampling

Model Tuning/Selection: Cross Validation, Walk Forward Estimation, AIC/BIC Criterions, Grid Search and Regularization

Time Series: ARIMA, Holt winters, Exponential smoothing, Bayesian structural time series

SQL: Subqueries, joins, DDL/DML statements.

PROFESSIONAL EXPERIENCE:

Confidential, Newark, NJ

Data Analyst

Responsibilities:

  • Analyzed data sources and requirements and business rules to perform logical and physical data modeling.
  • Analyzed and designed best fit logical and physical data models and relational database definitions using DB2. Generated reports of data definitions.
  • Involved in Normalization/De-normalization, Normal Form and database design methodology.
  • Maintained existing ETL procedures, fixed bugs and restored software to production environment.
  • Developed the code as per the client's requirements using SQL, PL/SQL and Data Warehousing concepts.
  • Involved in Dimensional modeling (Star Schema) of the Data warehouse and used Erwin to design the business process, dimensions and measured facts.
  • Worked with Data Warehouse Extract and load developers to design mappings for Data Capture, Staging, Cleansing, Loading, and Auditing.
  • Developed enterprise data model management process to manage multiple data models developed by different groups
  • Designed and created Data Marts as part of a data warehouse.
  • Wrote complex SQL queries for validating the data against different kinds of reports generated by Business Objects XIR2.
  • Using Erwin modeling tool, publishing of a data dictionary, review of the model and dictionary with subject matter experts and generation of data definition language.
  • Coordinated with DBA in implementing the Database changes and also updating Data Models with changes implemented in development, QA and Production. Worked Extensively with DBA and Reporting team for improving the Report Performance with the Use of appropriate indexes and Partitioning.
  • Developed Data Mapping, Transformation and Cleansing rules for the Master Data Management Architecture involved OLTP, ODS and OLAP.
  • Tuned and coded optimization using different techniques like dynamic SQL, dynamic cursors, and tuning SQL queries, writing generic procedures, functions and packages.
  • Experienced in GUI, Relational Database Management System(RDBMS), designing of OLAP system environment as well as Report Development.

Environment: ER Studio, Informatica Power Center 8.1/9.1, Power Connect/ Power exchange, Mainframes,DB2 MS SQL Server 2008, SQL,PL/SQL, XML, Windows NT 4.0, Tableau, Workday, SPSS, SAS, Business Objects, XML, Tableau.

Confidential, plrey park,CA

Data Analyst

Responsibilities:

  • Worked with internal architects, assisting in the development of current and target state data architectures.
  • Worked with project team representatives to ensure that logical and physical ER/Studio data models were developed in line with corporate standards and guidelines.
  • Involved in defining the business/transformation rules applied for sales and service data.
  • Implementation of Metadata Repository, Transformations, Maintaining Data Quality, Data Standards, Data Governance program, Scripts, Stored Procedures, triggers and execution of test plans
  • Define the list codes and code conversions between the source systems and the data mart.
  • Involved in defining the source to business rules, target data mappings, data definitions.
  • Responsible for defining the key identifiers for each mapping/interface.
  • Remain knowledgeable in all areas of business operations in order to identify systems needs and requirements.
  • Responsible for defining the key identifiers for each mapping/interface.
  • Performed data quality in Talend Open Studio.
  • Enterprise Metadata Library with any changes or updates.
  • Document data quality and traceability documents for each source interface.
  • Establish standards of procedures.
  • Coordinated meetings with vendors to define requirements and system interaction agreement documentation between client and vendor system.

Environment: Windows Enterprise Server 2000, SSRS, SSIS, Crystal Reports, DTS, SQL Profiler, and Query Analyze.

Confidential - Mentor, OH

Data Analyst

Responsibilities:

  • Business Reporting Requirements Analysis - Interacts with clients, participates in requirement gathering and system analysis
  • Wrote complex SQL queries to identify granularity issues and relationships between datasets and created recommended solutions based on analysis of the query results
  • Created reports and graphs using Business objects to meet the business requirements of the project. Worked closely with BA to determine feasibility and determine LOEs
  • Created and complete testing of repository, reports, dashboards and analytics to ensure they meet documented business requirements
  • Prepared Python / VBA scripts for data research, analysis and standardization.
  • Automated data report generation with SQL.
  • Automated Python scripts and used Tableau to perform ETL processes to load data large amounts of data into custom tracking tool used throughout Wholesale services.
  • Wrote the SQL queries on data staging tables and data warehouse tables to validate the data results.
  • Participated in user meetings, gathered Business requirements & specifications for the Data-warehouse design. Translated the user inputs into ETL design docs.
  • Created SQL queries and database triggers for state mandated reports and data elements.
  • Involved in the designing of the Data Model in deriving Conceptual, Logical and Physical model's using Erwin with Architect teams.
  • Defined, and documented the technical architecture of the Data Warehouse, including the physical components and their functionality.
  • Wrote complex SQL queries to identify granularity issues and relationships between datasets and created recommended solutions based on analysis of the query results
  • Data SME to the project team and business customers through documentation, drive-by questions, research, Visio, Excel pivot tables, and SQL queries
  • Designed ETL architecture to Process large no of files and created High-level design, low-level design documents.
  • Responsible for SQL tuning and optimization using Analyze, Explain Plan and optimizer hints.
  • Experience in creating UNIX scripts for file transfer and file manipulation.
  • Used VBA and macros to automate large Excel spreadsheets of 1M + rows of data for easier navigation, data scrubbing, conversion and manipulation.
  • Data mapping, logical data modeling, created class diagrams and ER diagrams and used SQL queries to filter data within the Oracle database.
  • Extracting the data from Azure Data Lake into HD Insight Cluster (INTELLIGENCE + ANALYTICS) and applying spark transformations & Actions and loading into HDFS.
  • Participated in Azure data Factory pipelines, datasets, copy and transform data in bulk via data Factory UI and Power Shell, scheduling and exporting data.
  • Customized and monitored incremental and full loading of Production using DAC. Worked with data architects team to make appropriate changes to the data models.
  • Configure DAC execution plans for scheduling to run ETL processes.

Environment: SQL, PL/SQL, Excel, ETL Ab initio, Microsoft Office Suite, Microsoft Visio. Files, Excel, Flat files, SQL Server

Confidential, Troy, Michigan

Data Analyst

Responsibilities:

  • Involved in defining the source to target data mappings, business rules, data definitions.
  • Involved in defining the business/transformation rules applied for sales and service data.
  • Worked with project team representatives to ensure that logical and physical ER/Studio data models were developed in line with corporate standards and guidelines.
  • Define the list codes and code conversions between the source systems and the data mart.
  • Coordinate with the business users in providing appropriate, effective and efficient way to design the new reporting needs based on the user with the existing functionality.
  • Worked with BTEQ to submit SQL statements, import and export data, and generate reports in Teradata.
  • Responsible for defining the key identifiers for each mapping/interface.
  • Implementation of Metadata Repository, Maintaining Data Quality, Data Cleanup procedures, Transformations, Data Standards, Data Governance program, Scripts, Stored Procedures, triggers and execution of test plans
  • Performed data quality in Talend Open Studio.
  • Coordinated meetings with vendors to define requirements and system interaction agreement documentation between client and vendor system.
  • Responsible for defining the functional requirement documents for each source to target interface.
  • Remain knowledgeable in all areas of business operations in order to identify systems needs and requirements.
  • Document the complete process flow to describe program development, logic, testing, and implementation, application integration, coding.
  • Enterprise Metadata Library with any changes or updates.
  • Generate weekly and monthly asset inventory reports.

Environment: SQL Server 2012/2008, Windows XP/NT/2000, MS-DTS, UML, UAT, SQL Loader, OOD, OLTP, PL/SQL, MS Visio.

Confidential

Data Analyst

Responsibilities:

  • Analyzed data sources and requirements and business rules to perform logical and physical data modeling.
  • Analyzed and designed best fit logical and physical data models and relational database definitions using DB2. Generated reports of data definitions.
  • Involved in Normalization/De-normalization, Normal Form and database design methodology.
  • Maintained existing ETL procedures, fixed bugs and restored software to production environment.
  • Developed the code as per the client's requirements using SQL, PL/SQL and Data Warehousing concepts.
  • Involved in Dimensional modeling (Star Schema) of the Data warehouse and used Erwin to design the business process, dimensions and measured facts.
  • Worked with Data Warehouse Extract and load developers to design mappings for Data Capture, Staging, Cleansing, Loading, and Auditing.
  • Developed enterprise data model management process to manage multiple data models developed by different groups
  • Designed and created Data Marts as part of a data warehouse.
  • Wrote complex SQL queries for validating the data against different kinds of reports generated by Business Objects XIR2.
  • Using Erwin modeling tool, publishing of a data dictionary, review of the model and dictionary with subject matter experts and generation of data definition language.
  • Coordinated with DBA in implementing the Database changes and also updating Data Models with changes implemented in development, QA and Production. Worked Extensively with DBA and Reporting team for improving the Report Performance with the Use of appropriate indexes and Partitioning.
  • Developed Data Mapping, Transformation and Cleansing rules for the Master Data Management Architecture involved OLTP, ODS and OLAP.
  • Tuned and coded optimization using different techniques like dynamic SQL, dynamic cursors, and tuning SQL queries, writing generic procedures, functions and packages.
  • Experienced in GUI, Relational Database Management System (RDBMS), designing of OLAP system environment as well as Report Development.
  • Extensively used SQL, T-SQL and PL/SQL to write stored procedures, functions, packages and triggers.
  • Analyzed of data report were prepared weekly, biweekly, monthly using MS Excel, SQL & UNIX.

Environment: SQL Server 2008R2/2005 Enterprise, SSRS, SSIS, Crystal Reports, Windows Enterprise Server 2000, DTS, SQL Profiler, and Query Analyzer.

Confidential

Data Analyst

Responsibilities:

  • Worked with internal architects, assisting in the development of current and target state data architectures.
  • Worked with project team representatives to ensure that logical and physical ER/Studio data models were developed in line with corporate standards and guidelines.
  • Involved in defining the business/transformation rules applied for sales and service data.
  • Implementation of Metadata Repository, Transformations, Maintaining DataQuality, DataStandards, Data Governance program, Scripts, Stored Procedures, triggers and execution of test plans
  • Define the list codes and code conversions between the source systems and the data mart.
  • Involved in defining the source to business rules, target data mappings, data definitions.
  • Responsible for defining the key identifiers for each mapping/interface.
  • Remain knowledgeable in all areas of business operations in order to identify systems needs and requirements.
  • Responsible for defining the key identifiers for each mapping/interface.
  • Performed data quality in Talend Open Studio.
  • Enterprise Metadata Library with any changes or updates.
  • Document data quality and traceability documents for each source interface.
  • Establish standards of procedures.
  • Coordinated meetings with vendors to define requirements and system interaction agreement documentation between client and vendor system.

Environment: SQL-Server 2008, Enterprise Architect, Power Designer, MS SSAS, Crystal Reports, SSRS, ER Studio, Lotus Notes, Windows XP, MS Excel, word and Access.

We'd love your feedback!