Data Analyst Resume
Denver, ColoradO
SUMMARY
- Data Modeler with 6 years of experience in data modelling, application support & maintenance for driving business solutions
- Professional working experience in Data Analytics using Microsoft SQL Server, SAS, Spark SQL, Power BI, Data Lake, Stream Analytics, Application Insights Analytics, Regression and Excel
- Advanced experience in Business Objects, SQL, and Microsoft Applications, including Excel and Access
- Experience in data mining using Spark, Hive SQL
- Experience with analytical data needs, handling of complex data requests, reports, and predictive data modeling
- As a Data Modeler experience working in Data Analysis, Data Profiling, Co - relating between Stovepipes data and confirming into the designed data model
- Working experience with Data Quality tools like IDQ
- Develop and maintain latest versions of data model, data dictionary, data lineage and other data architecture related design documents
- Extensively worked on ERWIN tool with features like Schema Generation, Complete Compare, Subject Area modeling etc.
- Good in Normalization / Demoralization techniques for effective and optimum performance in OLTP and OLAP environments.
- Having good experience with Normalization (1NF, 2NF and 3NF) and Denormalization techniques for improved database performance in OLTP, OLAP, Data Warehouse and Data Mart environments.
- Well-versed in designing Star and Snowflake Database schemas pertaining to relational and dimensional data modeling.
- Experience in design and implementation of enterprise-wide data strategy, data governance processes, data modeling best practices, data quality, data store, data security
- Deep understanding of Software Development Life Cycle (SDLC) as well as Agile/Scrum methodology to accelerate Software Development iteration
- Experience with analyzing online user behavior, Conversion Data (A/B Testing) and customer journeys, funnel analysis
- Having experience in writing complex SQL queries to perform end-to-end ETL validations and support Ad-hoc business requests. Also good in developed Stored Procedures, Triggers, Functions, Packages using SQL/PLSQL.
- Efficient in analyzing and documenting business and functional requirements along with Use Case Modeling and UML.
- Experience in data transformation, data mapping from source to target database schemas, data cleansing procedures.
- Hands on experience in writing queries in SQL and R to extract, transform and load (ETL) data from large datasets using Data Staging
- Experience in transating business requirements into working logical and physical data models
- Extensive experience in creating rich visualizations in dashboards using Tableau Dashboard and prepared user stories to create compelling dashboards to deliver actionable insights
TECHNICAL SKILLS
Core Competencies: Python, R(advanced), Java, SAS, Apache Spark, Hadoop, MySQL, Toad, Oracle DB, SQL, Excel(Advanced), Tableau, JMP Pro, Power BI, Rally, Informatica Erwin, Linux, Unix.
Statistical Methods: Binomial and Multinomial distributions, Univariate and Multivariate Statistics, Sampling, Hypothesis Testing, Confidence Intervals, Naïve Bayes, Likelihood functions, Probabilistic Classifiers etc.
Data Science: Regression, Clustering, Classification, Multidimensional Scaling Statistics for Data Science, Computing & Statistics, Applied Machine Learning, Advanced Business Analytics, Cloud Computing, Optimization Techniques, Project & Operations Management, Supply Chain Management, Social Media Marketing and Analysis.
PROFESSIONAL EXPERIENCE
Confidential - Denver, Colorado
Data Analyst
Responsibilities:
- Used SQL to extract and transform the data from Freud environment and load the structured data into the Smart Care environment. This reduced the service calls from
- Designed and developed various analytical reports from multiple data sources by blending data on a single Worksheet in Tableau Desktop to be presented to the CEO and CFO at Confidential
- Provided management with reorganization, consolidation and relocation strategies of the client facilities based on the Confidential t visit patterns which increased the services provided to facilities ratio using SSIS and SSRS
- Designed optimal appointment scheduling and staff planning for all the Confidential facilities by building regression model and using SAS, this helped in 10% reduction in the under allocated appointment hours
- Performed data analysis and profiling of source data to better understand the sources.
- Created 3 NF business area data modeling with de-normalized physical implementation; data and information requirements analysis.
- Redefined many attributes and relationships in the reverse engineered model and cleansed unwanted tables and columns as part of Data Analysis responsibilities.
- Performed legacy application data cleansing, data anomaly resolution and developed cleansing rule sets for ongoing cleansing and data synchronization.
- Conducted queries via Partners EHR/EMR system and output in SQL Server database as part of Readmission Project. Taking Hospital readmission dataset, after visualizing the patterns in Tableau and built a predictive model in R to predict readmission risk
- Developed logical data models and physical data models using ER-Studio
- Analyzed more than 100K Confidential t Records for early readmission risk using SAS, Py-Spark and Spark Machine Learning Library (MLlib)
- Developed algorithm to convert insurance-orientated ICD-9 codes to clinical practice meaningful disease classification using Python
- Implemented sentiment analysis of the tweets (mobile carriers) using NLTK sentiment analysis and twitter API
- Conducted data analysis using logistical model, KNN and random forest method to identify high readmission risk Confidential t and improved the accuracy (C-scores) by 30 percent
- Built complex SQL reports to audit $2.5 million of pay and insurance benefits for over 150 individual records
- Interacting with QA teams to review and develop data standards, data exchange, XML data standard, and data sharing model
- Involved in the planning phase of internal & external table schemas in Hive with appropriate static and dynamic portions for efficiency
Environment: SAS, SSIS, SSRS, Py-Spark, PL/ SQL, SQL Server, MS Visio, Python, R- Studio, Statistical Modelling, Power BI, Tableau, Google Analytics, Adobe Analytics, Rally
Confidential
Data Analyst
Responsibilities:
- Worked extensively on Erwin and ER Studio in several projects in both OLAP and OLTP applications
- Enabled the development of applications through the development of a cutting-edge logical data model that facilitated the creation of precise actuarial forecasting and analysis
- Designed ad-hoc queries with SQL in Congo’s Report Net. Examined reports and presented findings in PowerPoint and Excel
- Used Anomaly & Fraud detection techniques with SAS E-Miner for the Confidential client resulting in reduction of 22% of fraudulent cases
- Reporting of frauds, missed transactions, forecast, and user behavior using Tableau in direct weekly cross-functional team meetings to the Directors and the VP’s at AmEx for continuous process improvement
- Worked on ESB and SOA, Web Services methodologies using RESTful or SOAP based API’s
- Strong skills in Java, J2EE, Eclipse, Maven, Spring, Web Service Development and UML
- Implemented Agile Scrum practices for project implementation which reduced the project touch time by 300 man-hours and cost reduction of $30,000/year
- Headed an Analytics team of 10 for conducting statistical analysis to leverage the results to drive brand decision making and survey development
Environment: SAS, SQL, Erwin, Cognos, Tableau, Oracle 10i/11g, J2EE, Eclipse, Python, R, Maven, HTML, UML, TOAD, UNIX, Jira, Service Now
Systems Data Engineer
Confidential
Responsibilities:
- Developed the SQL table schema for the effective storage of the customer data
- Involved in preparing design and unit and Integration test documents
- Extensively involved in definition, design and implementation of complex actuarial calculations
- Involved in analyzing requirements and designing database using normalization techniques
- Created and maintained the Data Model repository as per company standards
- Evaluated performance of 300+ stores for Nielsen clients based on key metrics, performed statistical analysis, provided forecasts to enable stores to meet and exceed their financial targets through increased sales
- Worked on revenue data sets and devised a dynamic forecasting model through Regression Stepwise, KNN and Ensemble techniques. Average Ensemble model results are 92.3% accurate
- Created Data Lake by extracting customer’s data from various data sources into HDFS. This includes data from Teradata, Mainframes, RDBMS, CSV and Excel
- Involved in optimization of SQL scripts and designed efficient queries to query data
- Used Java, HTML, CSS, XML, FTP, and Batch scripting programming, in multiple projects based on requirement
Environment: Oracle 9i, SQL*Loader, Java, ODBC, TOAD, SQL, Excel, PowerPoint, Mainframe, HTML, CSS, FTP