We provide IT Staff Augmentation Services!

Data Analyst Resume

2.00/5 (Submit Your Rating)

Plano, TX

SUMMARY

  • 8 years of experience in Data Analysis, Data Migration, Data Cleansing, Transformation, Integration, Data Import, and Data Export using multiple ETL tools such as Ab Initio and Informatica Power Center Experience in testing and writing SQL and PL/SQL statements - Stored Procedures, Functions, Triggers and packages.
  • Experience on building Data Warehouse in platform, data bricks and data factory.
  • Have excellent knowledge on Python Collections and Multi-Threading.
  • Work with various Performed data migration into SQL Server, generated reports using SQL from oracle database, which were used for comparison with legacy system.
  • Good at extracting data from databases, Excel files using SAS/Base and SAS data steps.
  • Experience in Data Analysis, Data Modelling, Data Mining, Dashboard development and Testing.
  • Experience in Tableau report developing, testing and deploying reporting solutions using Tableau Server.
  • Working extensively with Dimensional Modeling, Data Migration, Data Transformation, Metadata, and Data dictionary, Data Loading, Modeling and Performance Tuning.
  • Proficient in designing of Star and Snowflake schemas with a very good understanding of fact and dimensional tables.
  • Skill experience in Python with proven expertise in using new tools and technical developments
  • Very good experience and knowledge on Amazon Web Services: AWS Redshift, AWS S3 and AWS EMR.
  • Experience in working on different Databases/Data warehouses like Teradata, Oracle, Apache, AWS Redshift, SQL Server.
  • Database software development using Oracle SQL & PL/SQL, Oracle Forms & Reports, Unix Shell Scripting and Java.
  • Expert in writing SQL queries and optimizing the queries in Oracle, SQL Server and Teradata.
  • Creating SQL scripts, PL/SQL programming, procedures, packages, functions in Oracle, SQL server, Teradata.
  • Experience in Programming Languages like Python, UNIX Shell Programming.
  • Work on MS SQL Server, Access Databases and developed T-SQL, SQL, Stored Procedures, Cursors, Triggers, Views, Indexes and Constraints, Query Optimizer
  • Use Data Analysis tools like Excel Spreadsheet, SQL objects, Business Intelligence Development Studio
  • Developed SQL, Stored Procedures, Views, Indexes and Constraints, Query Optimizer.
  • Strong Experience building software in Oracle SQL and PL/SQL utilizing all database objects like store procedure, functions, type objects, cursors, ref cursors, views, materialized views and PL/SQL collections.
  • Experience of importing various types of external data files into SAS dataset/SAS library, creating SAS datasets by using SAS/INFILE, SAS/IMPORT and SAS/SQL.
  • Work in collaborative team environment with other developers, business analysts, and a project management team. And highly motivated with the ability to work.
  • Experience in writing complex SQL Quires like Stored Procedures, triggers, joints, and Sub quires.
  • Expertise in SQL Server Storage Structures and Security Architecture for databases residing on storage.
  • Design & Implementation of Data Extraction, Transformation & Loading (using SQL), Analyze Oracle & SQL Server Data & Migration of the same.

TECHNICAL SKILLS

Tools: MS Office Suit (Word, Excel, Power Point), MS Project

AWS: S3, Redshift AWS, EC2

Tools: SAS, Base SAS, SAS/SQL, SAS/MACRO, SAS/ODS, SAS/ACCESS, Erwin, Pivot tables, Python, R, Snowflake schema, Star schema.

Programming Languages: SQL, PL/SQL, Python, R, UNIX Shell Programming

Databases: SQL Server, Oracle, MYSQL, PL/SQL, SQL, DB2

PROFESSIONAL EXPERIENCE

Confidential - Plano, TX

Data Analyst

Responsibilities:

  • Creating a multiple radar chart in R, supporting users to compare the performances.
  • Responsible for design and development of advanced Python programs to prepare to transform and harmonize data sets in preparation for Modeling.
  • Perform Data Analyst working on migrating data to the cloud (to Snowflake and AWS) from legacy data warehouses
  • Extract and analyze data from various sources - Data wrangling and cleanup using Python-pandas.
  • Work with Stake holders and Business Analysts to gather requirements and mapping these requirements to mapping document from source to destination.
  • Design ER diagrams, logical model (relationship, cardinality, attributes, and candidate keys) and convert these to physical data model including capacity planning, object creation and aggregation strategies, partition strategies, Purging strategies according to business requirements.
  • Perform Data Analysis on the Analytic data present in Teradata, AWS using SQL, Teradata SQL Assistant, and Python.
  • Created programs for Data Analysis using Python.
  • Using Tableau Desktop, created multiple rich dashboards visually telling stories of the business status, strength & weakness, potentials etc. for the client at a glance, also interact with data as necessary.
  • Perform Data Profiling and implementing Data Quality checks using Informatica developer, Python.
  • Responsible for technical Data governance, enterprise-wide Data modeling and Database design.
  • Perform Data mapping, logical data modeling, created class diagrams and ER diagrams and used SQL queries to filter data within the Oracle database.
  • Convert existing data archives to SAS databases to improve data quality and availability
  • Load the data from the panda’s data frames to the team's user defined space in Redshift database by using copy command from AWS S3 bucket.
  • Creating SQL queries in Toad to perform data analysis, data validation and data manipulation operations.
  • Analyze data through querying via SQL from various database sources (i.e. Oracle, SQl Server, MySQL)
  • Design and Develop Oracle PL/SQL and Shell Scripts, Data Import/Export, Data Conversions and Data Cleansing.
  • Use SQL Server Profiler to trace the slow running queries and tried to Optimize SQL queries for improved performance and availability.
  • Generate ad-hoc SQL queries using joins, database connections and transformation rules to fetch data from legacy Oracle and SQL Server database systems.
  • Create tables, stored procedures in SQL for data manipulation and retrieval, Database Modification using SQL, PL/SQL, Stored procedures, triggers, Views in MS SQL.
  • Implementing Star Schema methodologies in modeling and designing the logical data model into Dimensional Models.
  • Create the DDL scripts using ER Studio and source to target mappings (S2T- for ETL) to bring the data from multiple sources to the warehouse.
  • Analyze data using SAS for automation and determined business data trends.
  • Developing Python scripts to parse client data and shell scripts to manage and support software execution.
  • Working with Data Architect on Dimensional Model with both Star and Snowflake Schemas utilized.
  • Perform data manipulation on extracted data using Python.

Confidential, MN

Data Analyst

Responsibilities:

  • Engage user to gather software requirements rules and ensure alignment with development teams.
  • Assisted to build analytic tools to manage data and streamline data analyses using Python and SQL.
  • Complex incoming data cleaning and formatting using Python.
  • Ensured data accuracy through the creation and implementation of data integrity queries.
  • Prepared scripts to ensure proper data access, manipulation, and reporting functions with R programming languages.
  • Used tableau 10.x tool to achieve data report, analysis result visualization etc.
  • Provided technical assistance for development and execution of test plans and cases as per client requirements.
  • Assisted to build analytic tools to manage data and streamline data analyses using python and SQL.
  • Supported technical team members in development of automated processes for data extraction and analysis.
  • Participated in learning of techniques for statistical analysis projects, algorithms, and new methods.
  • Worked in AWS EC2 to maintain database and cronjob.

Confidential

Data analyst

Responsibilities:

  • Implemented Redshift based Data warehousing Solution
  • Automated GLUE ETL jobs using S3 events and AWS lambda functions
  • Pre-processed data for analysis by extracting data using complex SQL queries, joins, grouping and aggregation
  • Worked on POC’s using Spark, Python to load and transform data using data frames
  • Migrated historical data into AWS cloud and set up automated process in cloud for incoming Clickstream Data
  • Worked on Modeling and designing Database for data warehouse workload and created the objects in redshift
  • Well versed in data-preprocessing, to initialize the workflow and configure data cleaning process
  • Implemented Custom data extraction using SQL scripts, SQL stored procedures, Functions, Views and Triggers
  • Created and run jobs on AWS cloud to extract transform and load data into AWS Redshift using AWS Glue, S3 for data storage and AWS Lambda to trigger the jobs
  • Connected to AWS Redshift through Tableau to analyze the data for ad-hoc requirements as well as canned reports.
  • Build interactive dashboards in Tableau utilizing Parameters, Calculation fields and table calculations, Filters, Action filters and sets to handle views more efficiently
  • Working knowledge of AWS services models IaaS, PaaS, SaaS and other upcoming services in Cloud
  • Extensively used Agile methodology as the Organization Standard to implement the data Models.

Confidential

Data Analyst

Responsibilities:

  • Involved in analysis, design and documenting business requirements and data specifications. Supported data warehousing extraction programs, end-user reports and queries
  • Interacted with Business analysts to understand data requirements to ensure high quality data is provided to the customers
  • Worked on numerous ad-hoc data pulls for business analysis and monitoring by writing SQL scripts.
  • Created monthly and quarterly business monitoring reports by writing Teradata SQL queries includes System Calendars, Inner Joins and Outer Joins to retrieve data from multiple tables.
  • Developed BTEQ scripts in Unix using Putty and used cron-tab to automate the batch scripts and execute scheduled jobs in Unix
  • Performed verification and validation for accuracy of data in the monthly/quarterly reports.
  • Analyzed and validated data in Hadoop lake by querying through hive tables.
  • Created reports, charts by querying data using Hive Query Language and reported the gaps in lake data loaded.
  • Good knowledge on Json format data and performed the source, target validations using aggregations and null validity functions.
  • Created multi-set tables and volatile tables using existing tables and collected statistics on table to improve the performance.
  • Developed Teradata SQL scripts using RANK functions to improve the query performance while pulling the data from large tables.
  • Experience in performing Dual Data Validation on various Businesses critical reports working with another Analyst.
  • Designed stunning visualizations using tableau software and publishing and presenting dashboards on web and desktop platforms.
  • Designed and deployed reports with Drill Down, Drill Through and Drop-down menu option and Parameterized and Linked reports using Tableau.
  • Implement point of view security to Tableau dashboards to facilitate visibility across various levels of the Organization
  • Developed Python programs for manipulating the data reading from various Teradata and convert them as one CSV Files, update the Content in the database tables.

We'd love your feedback!