We provide IT Staff Augmentation Services!

Data Analyst Resume

Menlo Park, CA

SUMMARY

  • 10 + years of IT exp industry experience and 8+ years as a Data Analyst with a solid understanding of Data Modeling, Evaluating Data Sources, and a good understanding of Data Warehouse, and OLAP.
  • Knowledge in all phases of SDLC like Requirement Analysis, Implementation and Maintenance, and Good experience with Agile and Waterfall.
  • Good knowledge in a Python - based environment, along with data analytics, data wrangling, and Excel data extracts and Familiar with Pandas, NumPy, Matplotlib, SciPy, and PySpark.
  • Knowledge of maintaining and managing Tablaeu and Power BI drove reports and dashboards.
  • Practical understanding of the Data modeling (Dimensional & Relational) concepts like Star-Schema Modeling, Snowflake Schema Modeling, Fact, and Dimension tables.
  • Comprehensive knowledge and experience in process improvement, normalization/de-normalization, data extraction, data cleansing, data manipulation.
  • Good experience in data analysis, data profiling, data validation, developing low-level design patterns based on the business, and functional requirements.
  • Good experience in the Strategic development of a Data Warehouse and Performing Data Analysis and Data Mapping from an Operational Data Store to an Enterprise Data Warehouse.
  • Good knowledge in Metrics analysis, Reporting, Pivot tables, Charts, Graphs using (KPI) Key Performance Indicator.
  • Experience in RDBMS implementation and development using MySQL, SQL Server, SQL, and PL/SQL stored procedures and query optimization.
  • Experience in Data Visualization including producing tables, graphs, listings using various procedures and tools such as Tableau.
  • Good experience in extracting data from multiple sources to Power BI for Data Preparation.
  • Knowledge in RDBMS implementation and development using SQL, PL/SQL stored procedures and query optimization.
  • Exposure to different QA and development methodologies: Agile (Scrum), SDLC (Waterfall).
  • Experience in deriving data-driven insights and showcasing them using various business metrics and charts such as Regression, Time Series, and Histogram on reporting dashboards using Tableau and Data Studio..
  • Experience in cloud native, enterprises started discovering Infrastructure as a Service (IaaS) and Platform as a Service (PaaS) services and their utilization in building applications that leverage the elasticity and scalability of the cloud, all the while embracing the inherent failures of the cloud platform.
  • Experience in data Extraction, Transformation, and Loading ETL using various tools such as SQL Server Integration Services (SSIS) and Informatica.
  • Experience with Create Release Plan document using JIRA Dashboard reports, used Confluence for Release Notes, and shared release documents with Product Owner.
  • Good experience with databases like SQL, MySQL, SQL Server, and Teradata.
  • Experience in multiple operating systems including Linux and Windows.
  • Good communication skills, analytical mind, people skills, a team player, an individual contributor, and a quick learner to adapt to new technologies, processes, and business domains

TECHNICAL SKILLS

Programming Languages: SQL, R,Python, Statistical Analysis, Predictive Analysis

Database technologies: MySQL, MSSQL Server, PostgreSQL, Oracle SQL Developer, ETL Strategies

BI Tools: Tableau, Power BI Access, Pivot Tables, AWS(Redshift)

Visualization: Tableau, Power BI, Python, Matplotlib, Jupyter Notebook

Testing Tools: Win Runner, Load Runner, Test Director, Mercury Quality Center, Rational Clear Quest

PROFESSIONAL EXPERIENCE

Data Analyst

Confidential - Menlo Park, CA

Responsibilities:

  • Analyzed finance data for the billing and collections department to enhance revenue collection using SQL and MS Excel
  • Designed high level data pipelines to provide analysis of structured and unstructured data for major business initiatives
  • Collected and cleansed data to provide meaningful insights on the statistics of business performance
  • Analyzed historical data to identify annual and monthly trends in sales and revenue generation around the country
  • Cleaned and processed huge datasets using SQL Server and PostgreSQL databases to further generate reports and graphs for various departments
  • Created Pivot Tables by Excel and created visualization dashboards with interactive views in Tableau and public service for data reporting to supervisor and customers
  • Developed meaningful, efficient dashboards and visualizations using leading platforms, primarily Tableau.
  • Worked with other data and core business groups, such as Product, Sales, News and Engineering to help develop and enhance product databases and functionality
  • Extracted, interpreted, and analyzed million level data from multiple databases using SQL to identify key metrics, transforming raw data into meaningful, structured, actionable information using Python
  • Achieved automation of data cleaning, modeling and calculation for customer information database which helped in increasing efficiency of data flow pipelines
  • Delivered Interactive visualizations/dashboards using Tableau to present analysis outcomes in terms of patterns, anomalies, and predictions
  • Developed visualizations of the forecasted trend for profiling of data through Python and Tableau

Data Analyst/Data Scientist

Confidential - Raritan, NJ

Responsibilities:

  • Enhanced data extraction processes to include information that is relevant for building analytic systems, processing, cleansing, and verifying the integrity of data used for analysis & version control using Git and collaboration software such as Confluence, Jira, Bitbucket
  • Hands-on experience in working on building EDC and sandboxes for clinical data management and clinical trials related to drug investigation
  • Optimized ETL processes by automating python scripts and reducing the overall execution time by 25%
  • Analyzed data to find trends and behavior to drive business decisions using the insights reported from findings in reporting tools like Tableau to design interfaces using data collected from Medicare and Medicaid
  • Created visualization in Excel targeted to the non-technical audience to better understand and showcase the findings and insights for data collected from ERP healthcare software such as IQVIA
  • Understanding and writing SQL Scripts for development, automation of ETL process, error handling, and auditing purposes in SQL server
  • Built interactive dashboards and presented visualizations using BI tools like Tableau and statistical models with Python
  • Created action filters, parameters, and calculated sets for preparing dashboards and worksheets in Tableau
  • Developed a Chatbot using Rasa framework in Python for R&D at Confidential & Confidential to expedite the internal processing for healthcare research officers using NLP and text analytics techniques and deployed as a web-based application
  • Involved in analysis of Business requirement, Design and Development of High level and Low-level designs, Unit and Integration testing
  • Interacted with various other departments to understand and identify data needs and requirements to manage data retrieval accordingly from SQL and NoSQL databases such as Postgres, Mongo, MarkLogic for numerical and text data
  • Thorough grounding in all phases of data analysis, including definition and analysis of questions with respect to available data and resources, overview of data and assessment of data quality, selection of appropriate models and statistical tests and presentation of results
  • Data Mapping, logical data modeling, created class diagrams and ER diagrams and used SQL queries to filter data within the database along with data integration in
  • Strong experience in data extraction, manipulation, Analysis using claims data and reporting experience in the same using MS Excel and BI tools like Power BI, and python libraries like seaborn and matplotlib
  • Maintain a track of issues and bugs related to the product using JIRA, Confluence, Bitbucket & GitHub
  • Experience in analyzing the claims data, future trends, and summarized results to generate recommendations for data management using cloud services and databases - Oracle, MySQL and dashboarding tools like Tableau

Data Analyst

Confidential - NY

Responsibilities:

  • Involved in all phases of data acquisition, data collection, data cleaning, model development, model validation, and visualization to deliver data science solutions, designs and analytical programs
  • Worked closely with members services and financial department to analyze enrollment and billing related discrepancies
  • Processed and analyzed enrollment and claims portal data to detect and predict fraud claims
  • Supported the healthcare research department with analytical findings and need to optimize sales and effectively organize the data and report it using MS office tools
  • Worked on integrating data and deriving insights from it by comparing existing data with data collected from Medicare and Medicaid to improve current business scenario
  • Created classification models and interactive dashboards using BI and visualization tools to recognize web request with product association to classify the orders and scoring the products for analytics
  • Developed Tableau workbooks from multiple data sources using Data Blending and developed visualizations and dashboards using Tableau Desktop
  • Processed high volume data using SQL through Hadoop cluster and provided maintenance and support to the end clients to improve data and performance efficiency through Snowflake services
  • Customer segmentation based on their behavior or specific characteristics like age, region, income, geographical location and applying Clustering algorithms to group the customers based on their similar behavior patterns using statistical tools
  • Explored and managed data related to the management of patient eligibility processing, enrollment, claims and payments, analyzing it using SQL and Python as well as Hadoop, Hive and HiveQL
  • Created and maintained reports to display the status and performance of deployed model and algorithm with Tableau
  • Extensively used ETL methodology for supporting data extraction, transformations and loading processing, in a corporate-wide-ETL solution using ETL tools
  • Experience in designing stunning visualizations usingtableausoftware and publishing and presenting dashboards on web and desktop platforms and Proficient in MS Word,Excel.
  • Responsible for performing the data validation, process flow, dependency, Functionality Testing, a User Acceptance Testing and Database recovery and subsequent problem resolution
  • Data Mapping, Data Profiling, Data Quality, logical data modeling, created class diagrams and ER diagrams and used SQL queries to filter data within the database.
  • CreatedExceldocuments from project Bill of Materials/blueprints.

Data Analyst

Confidential - NY

Responsibilities:

  • Implemented Trajectory clustering methodologies to discover spatial and temporal traffic patterns in a traffic network
  • Identified spatially traffic flow groups using trajectory clustering and investigating temporal traffic patterns of each spatial group
  • Proposed use of Longest Common Subsequence (LCS) between two vehicles as their similarity measure
  • Incorporated LCS based-distance in trajectory clustering problem using density-based clustering algorithm (DBSCAN) using Pandas, Numpy, Matplotlib and Scikit libraries
  • Analyzed the output of clustering output in spatially distinct traffic stream clusters to provide an informative representation of major network traffic streams in ArcGIS
  • Pre-processed the network stream in ArcGIS to form a distance matrix to be used later to develop Cluster Representative Subsequence (CRS), reflecting dense road segments shared by trajectories belonging to a given traffic stream cluster
  • Generated CRS by merging pairwise LCSs via hierarchical agglomerative clustering and further classify trajectories and measure similarity between a new trajectory and cluster in Python

Junior Data Analyst

Confidential

Responsibilities:

  • Identifiedfraudin theinsurance claimwith the help ofunsupervised anomaly detection algorithmthat helped reduce the false alarm by 20% and increased the efficiency of the claim management system
  • Developed python script that regulated the approval of the payments and reserve of insurance claims under business constraints thereby increasing the customer database by 5%
  • MinedCustomer data forFarmers Insuranceand built alogistic regressionmodel inPythonto predict the probability of a customer buying auto insurance policy and achieved misclassification rate as 0.06
  • Useddata miningandclustering technique, to assign claims to most appropriate adjuster based on experience and loss type thus improving the overall customer experience in terms of claim duration
  • Leveraged on BI tools like Tableau and Power BI to develop business dashboards enabling leaders for decision makingandforecasting the number of claims monthly
  • PerformedData Exploration, Data Preparation, Data Cleaning and Data Visualizationof the raw Customer data for Farmers Insurance and builtclassificationmodel inPython to predict the probability of a customer buying auto insurance policy and improved the accuracy of the model by 13%
  • UsedData MiningandClustering technique, to assign claims to most appropriate adjuster based on experience and loss type thus improving the overall customer experience in terms of claim duration
  • Strong knowledge in implementing Waterfall and Agile (Sprint/Scrum) Methodology.
  • Active member in the Scrum team and involved in Scrum planning meetings, estimations and gathered information on requirements

Hire Now