Data Scientist Resume
Minneapolis, MinnesotA
SUMMARY:
- Above 7+ years of experience in large datasets of Structured and Unstructured data, Data Visualization, Data Acquisition, Predictive modeling, Data Validation.
- Extensive Experience in implementation functionalities like Grouping, Sorting, Derived Report parameters by using SSRS.
- Experience in writing expressions in SSRS and Expert in fine tuning the reports. Created many Drill through and Drill Down reports using SSRS.
- Collaborated with the lead Data Architect to model the Data warehouse in accordance to FSLDM subject areas, Snow flake schema, 3NF format.
- Experience in applying Predictive Modeling and Machine Learning algorithms for Analytical projects.
- Proficient in Statistical Modeling and Machine Learning techniques (Linear, Logistics, Decision Trees, Bayesian, XG Boost) in Forecasting/ Predictive Analytics, Regression based models, Hypothesis testing, Random Forest, SVM, K - Nearest Neighbors, Segmentation methodologies, Factor analysis/ PCA, Ensembles.
- Experience with Data Analytics, Graphs, Scales, Data Reporting, Ad-hoc Reporting, Pivot Tables and OLAP reporting.
- Experience in designing stunning visualizations using Tableau software and publishing and presenting dashboards, desktop platforms and Storyline on webs.
- Experience in coding SQL/PLSQL using Procedures, Triggers and Packages.
- Cluster Analysis, Principal Component Analysis (PCA), Recommender Systems, Association Rules.
- Highly skilled in using visualization tools like ggplot2, Tableau and d3.js for creating dashboards.
- Proficient in Predictive Modeling, Data Mining Methods, Factor Analysis, ANOVA, Hypothetical testing, normal distribution and other advanced statistical and econometric techniques.
- Proficiency in understanding statistical and other tools/languages - R, C, C++, Java, Python, SQL, UNIX, Qlikview data visualization tool and Anaplan forecasting tool.
- Experience in Deploying the SSIS Packages from development server to production server
- Experienced in using ETL tools in (SSIS) MS SQL 2016, 2014, MS SQL 2012, MS SQL 2008, MS SQL 2005 and DTS in MS SQL 2000.
TECHNOLOGIES:
SQL Server Tools: Enterprise Manager, SQL Server Profiler, SQL Server 2017/2012/2008/2005 Management Studio, DTS, SSIS, SSRS, SSAS, Performance Point Server 2010.
Utilities: Visual Studio, Visual Source Safe 6.0, Excel Power Pivot, Excel Data Explorer, DTS, Crystal Reports, Power Pivot, ProClarity, Microsoft Office.
Databases: MS SQL Server 2016/ 2014/ 2012 , MS-Access, Oracle 11g, Sybase and DB2.
Languages: C, C++, XML, HTML, DHTML, HTTP, PL/SQL, SQL, T-SQL,, Matlab.
Architecture: Relational DBMS, Client-Server Architecture, OLAP, OLTP, OLE-DB, XML, ASP, HTML, FTP.
Database Tools: logical data modeling, Kimball.
Operating Systems: Microsoft Windows 10/8/7/XP, UNIX and Linux.
PROFESSIONAL EXPERIENCE:
Confidential, Minneapolis, Minnesota
Data Scientist
Responsibilities:
- Responsible for design and development of advanced R-Python programs to prepare transform and harmonize data sets in preparation for modeling.
- Responsible for performing Machine-learning techniques regression-classification to predict the outcomes.
- Identifying and executing process improvements, hands on in various technologies such as Informatica and Business Objects.
- Developed and maintained data dictionary to create metadata reports for technical and business purpose.
- Involved in business process modeling using Unified Modelling Language.
- Interaction with Business Analyst and other Data Architects to understand Business needs and functionality for various project solutions.
- Handled importing data from various data sources, performed transformations using Hive, Map Reduce, and loaded data into HDFS.
- Involved with Data Analysis primarily Identifying Data Sets, Source Data, Source Meta Data, Data Formats and Data Definitions.
- Created SQL tables with referential integrity and developed queries using SQL, PLSQL and SQL*PLUS.
- Created PL/SQL packages, procedures, functions, Triggers.
- Collaborate the data mapping document from source to target and the data quality assessments for the source data.
- Prepare ETL architect& design document which covers ETL architect, extraction, SSIS design, transformation and loading of Duck Creek data into dimensional model.
- Participated in Business meetings to understand the business needs and requirements.
- Provide technical & requirement guidance to team members for ETL -SSISdesign.
- Design Logical and Physical Data Model using MS Visio 2003 data modeler tool.
- Coordinate & communicate with technical teams for any data requirements
- Design ETL framework & development.
- Participated in Architect solution meetings and guidance in Dimensional Data Modeling design.
- Participated in stake holders meetings to understand the business needs and requirements.
Environment: . Spark MLlib, TensorFlow, Python, K- means, ANN, Regression, Accord.NET, Oryx 2, Flask,, Naive Bayes, SVM, Amazon Machine Learning (AML), ORM, Jinja 2, Apache,Django, Mako.
Confidential, Findlay, Ohio
Data Scientist
Responsibilities:
- Document the complete process flow to describe program development, testing, application integration, coding & implementation.
- Data modeling & formulation of statistical equations using advanced statistical forecasting techniques.
- Built predictive scorecards for Life Insurance, TD, Cross-selling Car loan & RD.
- Scoring predictive models as per regulatory requirements & ensuring deliverables with PSI.
- Developing propensity models for Retail liability products to drive proactive campaigns.
- Mentoring Provide guidance to team members.
- Responsible for defining the functional requirement documents for each source to target interface.
- Transformation, Data cleansing & creating new variables using R.
- Approve & Present designed Logical Data Model in Data Model Governance Committee (DMGC)
- Identifying the Customer & account attributes required for MDM implementation from disparate sources & preparing detailed documentation.
- Validated the machine learning classifiers using ROC Curves & Lift Charts.
- Work with users to identify the most appropriate source of record & profile the data required for sales & service.
- Tabulation & Extraction of data from multiple data sources using R, SAS.
- Responsible for defining the key identifiers for each mapping/interface.
- Extracted data from HDFS & prepared data for exploratory analysis using datamunging.
- Arrange & chair Data Workshops with SME's & related stake holders for requirement data catalogue understanding.
Environment: Hadoop, logistic regression, regression, forest, OLAP, HDFS, NLTK, SVM, JSON and XML, Teradata, OLTP, Unix, Python, MLLib and SAS.
Confidential, Cincinnati, OH
Data Analyst
Responsibilities:
- Deployed GUI pages by using HTML, JSP, JSTL, DHTML, XHTML, CSS, JavaScript, AJAX
- Implemented the online application by using Core Java, Jdbc, JSP, Servlets & EJB 1.1, Web Services, SOAP, WSDL
- Used SAX & DOM parsers to parse the raw XML documents
- Used Singleton, factory design pattern, DAO Design Patterns based on the application requirements
- Preparing & executing Unit test cases
- Used RAD as Development IDE for web applications.
- Involved in fixing bugs & minor enhancements for the front-end modules.
- Used Log4J logging framework to write Log messages with various levels.
- Doing functional & technical reviews
- Implemented Microsoft Visio & Rational Rose for designing the Use Case Diagrams, Class model, Sequence diagrams & Activity diagrams for SDLC process of the application
- Guaranteeing quality in the deliverables.
- Implemented the project in Linux environment
- Maintenance in the testing team for System testing/Integration/UAT
- Was a part of the complete life cycle of the project from the requirements to the production support
- Conducted Design reviews & Technical reviews with other project stakeholders.
- Created test plan documents for all back-end database modules
Environment: JAVA, R 3.0, Erwin 9.5, Tableau 8.0, MDM, QlikView, MLLib, PL/SQL, HDFS, Teradata 14.1, JSON, MapReduce, PIG, Spark, R Studio, MAHOUT.
Confidential
Data Scientist
Responsibilities:
- Created Web Pages using XML, XSLT, JSP, HTML and JavaScript.
- Based on business requirement, developed the SQL queries, Stored procedure, Views, Trigger to implementing the business rules and transformations.
- Involved in requirements gathering and preparing software requirement specification.
- Optimized the code and database for maximum performance.
- Responsible for deploying Web Logic application server.
- Design and maintain the database to support the application using Oracle8i.
- Have strong sql coding knowledge.
- Prepared the complete data mapping for all the migrated jobs.
- Data Migration from Flat files, MS-Access, Excel, SQL Database.
- Exposure to the UNIX shell scripting tasks.
- Involved in Query optimization, performance tuning and index creation.
- Prepared Documentation of the system. Writing JUnit Test cases for testing the application
- Created Stored Procedure and Triggers to enforce business rules.
Environment: HTML, CSS, javascript, XML, Java, J2ee, Servlets, JSP, EJB, MVC, SQL Server, SQL Assistant, WIN2000/XP, XSLT, JUnit, MS Access, UML, Windows.
Confidential
Data Scientist
Responsibilities:
- Developed the User Interactive web pages in a professional manner with using web technologies like HTML, XHTML, and CSS as per company’s standards.
- Involved in development, design and implementation of front end part widget based application.
- Developed Web forms/user interfaces using Struts MVC, jQuery and JSP.
- Developed presentation-tier JSP pages in HTML, Implemented jQuery Data Grid control, Validation control and other Widget controls.
- Worked on AJAX controls like Update Panel to manage the post back of the web page to server, AJAX Script Manager and Script Manager Proxy controls to register JavaScript and web service files.
- Used Ajax Controls, Web forms, JavaScript and HTML for Commission, Payment and Inventory reports.
- Developed various AJAX controls and widgets to build a rich User Interface for the application.
- Used JavaScript accordingly for validation purpose, browser detection and controls.
- Used Cascading Style Sheets (CSS) to maintain design consistency across all web forms.
- Worked with XPath, XML Node, XML Data documents to synchronize with dataset
- Created Windows services to create automation processing of XML.
- Created Model objects using Entity Data Model with Entity framework using Hibernate.
- Involved in developing jQuery and AJAX wrapper classes for fast retrieval of data and for animations.
Environment: HTML, JavaScript, CSS, jQuery, XML, XPath, AJAX, JSP, MVC, Junit, Window OS.