Data Scientist Resume
OH
SUMMARY:
- Around 7 years of experience in Data science with large Data Sets of Structured and Unstructured Data, Data Acquisition, Data Validation, Predictive modeling, Data Visualization, Web Crawling, Web Scraping Statistical Modeling, Data Mining and Natural Language Processing (NLP).
- Good in and different methodologies like Waterfall, Agile and Scrum.
- Developed intricate algorithms based on deep - dive Statistical Analysis and Predictive Data Modeling that were used to deepen relationships, strengthen longevity and personalize interactions with customers.
- Experience in Data Mining, Machine learning and Spark Development with big informational datasets of Structured and Unstructured information, Data Acquisition, Data Validation, Predictive Demonstrating, Data Visualization.
- Extensive experience working with Django, Flask frameworks.
- Experienced in Python data manipulation for loading and extraction as well as with python libraries such as matplotlib, NumPy, SciPy and Pandas for data analysis.
- Experience with Statistics, Data Analysis, Machine Learning using Python and R Language.
- Worked with complex applications using R, SPSS and Python to develop neural network algorithms, cluster analysis.
- Proficient in Tableau data visualization tool to analyze and obtain insights into large datasets, create visually powerful and actionable interactive reports and dashboards.
- Experience in analyzing Format data using Machine Learning algorithm by Python Scikit-Learn.
- Strong experience working with databases like SQL Server, Oracle, MS Access, No SQL databases like MongoDB and Cassandra.
- Experience in data mining, including predictive behavior analysis, Optimization and Customer Segmentation analysis using SAS and SQL
- Experienced in developing web-based applications using Python, Django, XML, CSS3, HTML5, JavaScript and jQuery.
- Extensively worked on data warehousing and decision support systems with relational databases such as Oracle design and database development using SQL, PL/SQL, SQL PLUS and Toad.
- Expertise in SQL database testing involving SQL Queries, joins to pull data from database.
- Excellent analytical, problem solving, communication and interpersonal skills, with ability to interact with individuals at all levels.
TECHNICAL SKILLS:
Development methodology: SDLC, Agile, Scrum, Waterfall
Framework: Django, Flask frameworks.
Programming Languages: Python, Django, XML, CSS3, HTML5, JavaScript, R, SPSS, SAS, Scikit-Learn, Tableau, jQuery, JSON, AJAX, R Studio, Machine learning
Library & Packages: matplotlib, NumPy, SciPy, Pandas, MLLIB, MLX tend, NLTK
Database/Database tools: SQL Server, Oracle, MySQL, Toad, PL/SQL
PROFESSIONAL EXPERIENCE:
Confidential, OH
Data Scientist
Responsibilities:
- Implemented Agile Methodology for building an internal application.
- Interpret and communicate findings from analysis to engineers, product and stakeholders.
- Application of various algorithms and statistical modeling like decision trees, regression models, neural networks, SVM, clustering to identify Volume using scikit-learn package in python, MATLAB.
- Create and publish multiple dashboards and reports using Tableau server.
- Worked on Data audit of SAS code/projects and sense check of results.
- Worked on different data formats such as JSON, XML, and performed machine learning algorithms in Python.
- Programmed a utility in Python that used multiple packages (SciPy, NumPy, and Pandas and R (data. table, quant mod, ggplot2).
- Performed analysis using predictive modeling, data/text mining, and statistical tools.
- Used Machine Learning Linear regression models, KNN, and K-means clustering algorithms.
- Worked with migrating the Django database from MySQL to PostgreSQL
- Spark implementation for using python, spark SQL to access hive tables into spark for faster processing of data.
- Created Data Quality Scripts using SQL and Hive to validate successful data load and quality of the data.
- R programming language for graphically critiquing the data and performed data mining. Interpreting Business requirements, data mapping specifications and responsible for extracting data as per the business requirements.
- Worked on No SQL databases including Cassandra, Mongo DB, Mark Logic and HBase to access the advantages and disadvantages of them for a particular goal of a project.
Environment: Agile, python, MATLAB, neural networks, SVM, scikit-learn, SAS, JSON, XML, SciPy, NumPy, pandas, Django, KNN, MySQL, R programming
Confidential, OH
Data Scientist
Responsibilities:
- Implemented Waterfall Methodology for building an internal application and Software Development Life Cycle (SDLC), Defect lifecycle management.
- Performed data analysis and profiling of source data to better understand the sources.
- Data Cleaning, features scaling, features engineering using pandas and NumPy packages in python.
- Designed, developed and maintained daily and monthly summary, trending and benchmark reports repository in Tableau Desktop.
- Designed & developed various Ad hoc reports for different teams in Business (Teradata and MSACCESS, MSEXCEL).
- Involved in creating dashboards and reports in Tableau 8.1. Created report schedules on Tableau server.
- Worked utility in Python that used multiple package.
- Data visualization, reporting using SSRS.
- Worked With different data formats such as JSON, XML and performed machine learning algorithms in Python.
- Design and implemented algorithms to detect and respond to faulty inputs.
- Worked with maintaining using Spark SQL.
- Extracting data from the database using ETL Concept.
- Designed & developed various Ad hoc reports for different teams in MS EXCEL.
- Used SAS, SQL, Oracle and MS Office analysis tools to complete analysis requirement.
- Generated weekly and monthly reports and maintained, manipulated data using SAS Tableau.
Environment: SDLC, Waterfall, NumPy, python, Teradata, MSACCESS, MSEXCEL, SSRS, JSON, XML, Spark SQL, ETL, MS EXCEL, Oracle, Tableau
Confidential
Data Scientist
Responsibilities:
- Applied the Agile and (SLDC) software development process to establish a business analysis methodology.
- Interpret and communicate findings from analysis to engineers, product and stakeholders.
- Identify meaningful insights from chargeback data. created various types of data visualizations using R, and Tableau.
- Used analytics libraries MLLIB and MLX tend.
- Experiences in Data Visualization using matplotlib.
- Worked with migrating the Django database from MySQL to PostgreSQL.
- Knowledge in Python's multiple data science packages like NumPy, Seaborn, SciPy and NLTK.
- Knowledge in large data read/writes to and from csv and Excel files using pandas.
- Communicate and coordinate with other departments to collection business requirement.
- Develop a technical brief based on the business brief.
- Worked on Data audit of SAS code/projects and sense check of results.
- Create and publish multiple dashboards and reports using Tableau server.
- Responsible for maintaining and analyzing large datasets used to analyze risk by domain experts.
- Develop Hive queries that compared new incoming data against historic data.
Environment: Agile, SLDC, R, Tableau, Data Visualization, matplotlib, NumPy, pandas., Seaborn, SciPy, NLTK, Django, MySQL, PostgreSQL, csv, SAS, Tableau
Confidential
Data Scientist
Responsibilities:
- Conducted research on development and designing of sample methodologies Like waterfall, Scrum, analyzed data regarding the claim status, coverages and estimate cost of medical procedures.
- Developed Machine Learning algorithms to find out the number of claims that are properly being claimed and suggest customers by providing insights for the smarter healthcare.
- Generated graphs and reports using matplotlib, and pandas’ packages in python for analytical models.
- Worked with python libraries like matplotlib, Python and NumPy.
- Developed and implemented R application which showcases machine learning for business forecasting.
- Experience working in project with machine learning, big data, data visualization, R and Python development, Unix, SQL.
- Developed customized Tableau visualizations and stories using tableau desktop.
- Worked on databases like MS SQL server, PostgreSQL and SQLite.
- Built predictive models to forecast residual values of leased vehicles and estimated the demand for these used vehicles using Regression and Time Series techniques in SAS.
- Programming experience with SQL including relational database query construction, optimization and predictive modeling.
- Generated graphs and reports using ggplot package in R Studio for analytical models.
Environment: waterfall, Scrum, matplotlib, pandas, Python, R, Tableau, Unix, SQL., MS SQL server, PostgreSQL, SAS, R Studio