We provide IT Staff Augmentation Services!

Data Analyst Resume

5.00/5 (Submit Your Rating)

Morrisville, PA

SUMMARY

  • Over 8+ years of strong experience in Business and Data Modeling/ Data Analysis, Data Architect, Data Profiling, Data Migration, Data Conversion, Data Quality, Data Governance, Data Integration, MDM, NoSQL and Metadata Management Services and Configuration Management.
  • Expertise and Vast knowledge of Enterprise Data Warehousing including Data Modeling, Data Architecture, Data Integration (ETL/ELT) and Business Intelligence.
  • Experienced in using various Teradata Utilities like Teradata Parallel Transporter (TPT), Mload, BTEQ, FastExport, and Fastload.
  • Experienced in Dimensional Data Modeling experience using Data modeling, Relational Data modeling, ER/Studio, Erwin, andSybase PowerDesigner, Star Join Schema/Snowflake modeling, FACT& Dimensions tables, Conceptual, Physical & logical data modeling.
  • Extensive experience in development of T - SQL, OLAP, PL/SQL, Stored Procedures, Triggers, Functions, Packages, performance tuning and optimization for business logic implementation.
  • Proficient in handling complex processes using SAS/ Base, SAS/ SQL, SAS/ STAT SAS/Graph, Merge, Join and Set statements, SAS/ ODS.
  • Experience in Creating Audit control system for ETL process for Big Data and Data Warehouse Application.
  • Experience in developing Map Reduce Programs using ApacheHadoop for analyzing the big data as per the requirement.
  • In depth knowledge of software development life cycle (SDLC), Waterfall, Iterative and Incremental, RUP, evolutionary prototyping and Agile/Scrum methodologies, as well as data warehouse concepts such as OLTP, OLAP, ETL, Star and Snowflake Schema, data mapping, facts, and dimensions.
  • Excellent understanding of Microsoft BI toolset including Excel, Power BI, SQL Server AnalysisServices, Visio, Access.
  • Sound knowledge of SDLC process - Involved in all phases of Software Development Life Cycle - analysis, design, development, testing, implementation, and maintenance of applications.
  • Performing application level DBA activities creating tables, monitored and tuned TeradataBETQ scripts using Teradata Explain utility.
  • Thorough Knowledge in creating DDL, DML and Transaction queries in SQL for Oracle and Teradata databases.
  • Technically proficient, Customer dedicated with remarkable experience in Mainframe development & Maintenance projects built with Teradata, JCL & IBM Tools.
  • Experienced in Creating and Maintaining documentation such as Data Architecture/Technical/ETL Specifications.
  • Well versed in 3NF and Star Schema(Dimensional) Data Modeling techniques.
  • Experience in Handling Huge volume of data in/out from Teradata/Big Data.
  • Worked and extracted data from various database sources like Oracle, SQL Server, DB2, and Teradata.
  • Hands on experience with modeling using ERWIN in both forward and reverse engineering cases.
  • Strong exposure to writing simple and complex SQL, PL/SQL queries.
  • Experienced as Data Analyst. Have hands-on experience in Onsite-Offshore model projects.
  • Involved in writing shell scripts on UNIX for Teradata ETL tool and data validation.

TECHNICAL SKILLS

Tools: Tableau, Spotfire, Informatica 9.6.0, Microstatergy, Test Management Tool (QC 9.2, ALM 11),SVN, Shiny with R

Languages: SQL, R, Python, C, Esqlc, UNIX shell scripting

Database: SQL, Hive, Databases SQL-Server, My SQL, MS Access, Teradata, Netezza, Mongo DB.

Data Science: Predictive Analytics, Fraud Detection, Pattern Mining, Sentiment Analysis, Anomaly Detection, Outlier Detection

Machine Learning Algorithm: Classification (Random Forest,K-nearest neighbor,NaivesBayes,AdaBoost,SVM ) Clustering (K-Means,Hierarchical clustering), Regression - (Linear Regression,Logistic Regression)

Analytics: Python (NLTK, NLP, Scikit-learn, Numpy, Panda, SciPy, plotly), R

Data Engineering: Data Mining (Python, R, SQL)

BI Tools: Tableau, Tableau server, Tableau Reader, SAP Business Objects, OBIEE, QlikView, SAP Business Intelligence, Amazon Redshift, or Azure Data Warehouse

ETL Tools: Informatica PowerCenter 7.x/ 8.x/ 9.x, Talend ETL, Pentaho, ODI

Operating Systems: Ubuntu, Windows.

PROFESSIONAL EXPERIENCE

Data Analyst

Confidential, Morrisville, PA

Responsibilities:

  • Wrote several Teradata SQL Queries using Teradata SQL Assistant for Ad Hoc Data Pull request.
  • Developed Python programs for manipulating the data reading from various Teradata and convert them to one CSV Files.
  • Performing statistical data analysis and data visualization using Python and R
  • Worked on creating filters, parameters and calculated sets for preparing dashboards and worksheets in Tableau.
  • Created new scripts for Splunk scripted input for collecting CPU, system and OS data.
  • Interacting with other data scientists and architected custom solutions for data visualization using tools like a tableau, Packages in R and R-Shiny.
  • Maintenance of large data sets, combining data from various sources by Excel, SAS Grid, Enterprise, Access and SQL queries.
  • Analysed DataSet with SAS programming, R, and Excel.
  • Publish Interactive dashboards and schedule auto- data refreshes
  • Experience in performing Tableau administering by using tableau admin commands.
  • Developed normalized Logical and Physical database models for designing an OLTP application.
  • Knowledgeable in AWS Environment for loading data files from on prim to Redshift cluster
  • Performed SQL Testing on AWS Redshift databases
  • Developed TeradataSQL scripts using OLAP functions like rank and rank () Over to improve the query performance while pulling the data from large tables.
  • Involved in running Map Reduce jobs by processing millions of records.
  • Written complex SQL queries using joins and OLAP functions like CSUM, Count, and Rank etc.
  • Experienced in migrating HiveQL into Impala to minimize query response time.
  • Responsible for Data Modeling as per our requirement in HBase and for managing and scheduling Jobs on a Hadoopcluster using Oozie jobs.
  • Worked on Spark SQL and Data frames for faster execution of Hive queries using Spark SQL Context.
  • Design and development of ETL processes using Informatica ETL tool for dimension and fact file creation.
  • Develop and automate solutions for a new billing and membership Enterprise data Warehouse including ETL routines, tables, maps, materialized views, and stored procedures incorporating Informatica and Oracle PL/SQL toolsets.
  • Performed analysis of implementing Spark using Scala and wrote spark sample programs using PySpark.
  • Created UDF to calculate the pending payment for the given residential or small business customer's quotation data and used in Pig and Hive Scripts.
  • Experienced in moving data from Hive tables into HBase for real-time analytics on Hive tables.
  • Handled importing of data from various data sources, performed transformations using Hive. (External tables, partitioning)

Environment: SQL/Server, Oracle 9i, MS-Office, Teradata, Informatica, ER Studio, XML, Hive, HDFS, Flume, Sqoop, R connector, Python, R, Tableau.

Data Analyst

Confidential, Dublin, Ohio

Responsibilities:

  • Worked on Requirement Analysis, Data Analysis and Gap Analysis of various source systems sitting and coming from multi-systems. Responsible for BI Data Quality.
  • Conducted JAD sessions to allow different stakeholders such as editorials, designers, etc.,
  • Performed Business Process mapping for new requirements.
  • Performed Data Validationwith Data profiling
  • Designed reports in Access, Excel using advanced functions not limited to VLOOKUP, pivot tables, formulas
  • Use SQL, PL/SQL to validate the Data going into the Data Ware House
  • Conducted UAT (User Acceptance Testing) for multiple iterations by writing Test Cases and signed off the same after approval
  • Written XSLT scripts for various XML needs and used XPath, XQuery for data retrieval from XML documents.
  • Responsible for creating the Requirements Traceability Matrix.
  • Involved in Designing Star Schema, Creating Fact tables, Dimension tables and defining the relationship between them.
  • Tested the ETL Informatica mappings and other ETLProcesses ( Data Warehouse Testing)
  • Verified and maintained Data Quality, Integrity, data completeness, ETL rules, business logic.
  • Tested several UNIX Korn Shell scripts for ETLdata loading
  • Verified Informatica sessions, worklets, and workflows in QA repository.
  • Extensively used SQL queries to check storage and accuracy of data in database tables.
  • Used SQL for Querying the Oracle database.
  • Performed all aspects of verification, validation including functional, structural, regression, load and system testing.
  • Involved in backend testing for the front end of the application using SQL Queries in Teradata database.
  • Tested whether the reports developed in Cognos are as per company standards.
  • Tested different detail, summary reports and on-demand reports using Report Studio.
  • Reported bugs and tracked defects using Quality Center 10.
  • Experienced in working with DB2, Teradata

Environment: Oracle 11g, SQL Server 2005, Informatica Power Center 9/8.5.1 (Workflow Manager, Workflow Monitor), SSAS, SAS/Stat, SAS/Graph, SAS/ODS, SAS/SQL, SAS/Connect, SAS/Access, Cognos 8.0 Series, Netezza, XML, Test Cases, Assistant, Teradata V2R6, Teradata SQL Assistant, IBM DB2, Unix, Test Matrix, TOAD, HP Quality Center 10.

Data Analyst

Confidential -Burlington, VT

Responsibilities:

  • Worked with the development and testing teams about the ETL processes involved in the project.
  • Worked on Requirement Analysis, Data Analysis and Gap Analysis of various source systems sitting and coming from multi-systems. Responsible for BI Data Quality.
  • Extracted Data using SSIS from DB2, XML, Oracle and flat files, Excel perform transformations and populate the data warehouse
  • Expert hand on knowledge of SAS/BASE, SAS/MACRO, SAS/STAT, SAS/GRAPH, SAS/ACCESS.
  • Performed Teradata SQL Queries, Creating Tables, and Views by following Teradata Best Practices.
  • Conducted JAD sessions to allow different stakeholders such as editorials, designers, etc.,
  • Performed Business Process mapping for new requirements.
  • Use SQL, PL/SQL to validate the Data going into the Data Ware House
  • Creating complex data analysis queries to troubleshoot issues reported by users
  • Evaluates data mining request requirements and help develop the queries for the requests.
  • Worked with Business Analysts, developers while reviewing the Business Requirement Documents and when there are enhancements in the applications.
  • Defined data requirements and elements used in XML transactions.
  • Developed Test Cases for Deployment Verification, ETLData Validation, Cube Testing and Report testing.
  • Tested the ETL process for both before data validation and after data validation process. Tested the messages published by ETL tool and data loaded into various databases
  • Independently designed and developed project document templates based on SDLC methodology.
  • Prepared Entry and Exit criteria for Testing cycles, review with functional leads and PMO.
  • Exported Requirements, test plans and test cases to QC 9.2.
  • Created Traceability Matrix to ensure that all requirements are covered in test cases.
  • Prepared Kick-off meeting PowerPoint and gave a presentation on testing activities for each cycle of testing and for different releases.
  • Wrote complex SQL, PL/SQL Testingscripts for Backend Testing of the datawarehouse application. Expert in writing Complex SQL/PLSQL Scripts in querying Teradata and Oracle.
  • Used TOADSoftware for Querying ORACLE and Used WinSQL for Querying DB2.
  • Analysed business requirements, system requirements, data mapping requirement specifications, and responsible for documenting Testcases, Test Scripts in Quality Center 9.2
  • Created defects, manage defect reports and defect status in ClearQuest tool
  • Extensively tested the BusinessObjects report by running the SQL queries on the database by reviewing the report requirement documentation.

Environment: Informatica Power Center 8.6(Repository Manger, Designer, Server Manager), Oracle 10g, MS SQL, HP Quality Center 9.2, Netezza, Requirement traceability Matrix (RTM), XML, Assistant, Teradata V2R6, Teradata SQL Assistant, IBM DB2, SSAS, SAS/Stat, SAS/Graph, SAS/ODS, SAS/SQL, SAS/Connect, SAS/Access, IBM Rational ClearQuest 7.0, TOAD, Business Objects, MS Excel, MS Project.

Data Analyst

Confidential -Baton Rouge, LA

Responsibilities:

  • Analyze the client data and business terms from a data quality and integrity perspective.
  • Perform root cause analysis on smaller self-contained data analysis tasks that are related to assigned data processes.
  • Involved in Designing Star Schema, Creating Fact tables, Dimension tables and defining the relationship between them.
  • Analyzed business requirements, system requirements, data mapping requirement specifications, and responsible for documenting Test cases, Test Scripts in Quality Center 9.2
  • Written XSLT scripts for various XML needs and used XPath, XQuery for data retrieval from XML documents.
  • Tested the ETL Informatica mappings and other ETL Processes ( Data Warehouse Testing)
  • Verified and maintained Data Quality, Integrity, data completeness, ETL rules, business logic.
  • Developed numerous Teradata SQL Queries by creating SET or MULTISET Tables, Views, Volatile Tables, using Inner and Outer Joins
  • Implemented Indexes, collecting Statistics and Constraints while creating Tables and Views.
  • Resolving Data issues to improve data quality.
  • Deployed scripts using Bash for automating daily tasks, improve security wide VPC network and produce daily SQL dumps and backup of confidential data to AWS S3/Glacier.
  • Developed complex SQL queries to bring data together from various systems.
  • Performed Dataalignment and Data Cleansing. Involved in Data Migration between Teradata, MS SQLServer,and Oracle.
  • Archived the historical datasets on Teradata to UNIX server using SASscripts and created them back to Teradata whenever required.
  • Sourced and analyzed data from a variety of sources like SASdatasets, MS Access, MS Excel, CSV and flat files etc.
  • Assisted the team for standardization of reports using SAS macros and SQL.
  • Create Teradata tables, Views, Macros and analyzed various Teradata tables for UPI's and Monitor the same. Execution of MainframesJCL scripts.
  • Utilized ODBC for connectivity to Teradata &MS Excel.

Environment: Agile, XML, SQL, Root Cause Analysis, Requirement Traceability Matrix (RTM), XSLT, Informatica, ETL, Teradata, AWS S3 / Glacier, MS SQL, Oracle, UNIX, SAS, MS Access, MS Excel, CSV, Flat files, SAS Macros, JCL, ODBC, Unix Shell, UAT.

Data Analyst

Confidential

Responsibilities:

  • Contributed in creating cost analysis infrastructures for profitable trade and asset evaluation, as a part of asset management division in the banking and financial domain
  • Developed ETL data pipelines from OLTP databases to OLAPDataWarehouse using SSIS providing data migration solutions
  • Performed data integration and transformation on data sources with 100,000+ records for upstream analysis and reporting
  • Created reusable packages to extract data from multi formatted files and modified existing SQL queries to optimize the data performance
  • Monitored daily incidents raised in service now under Incident Management Process (ITIL) and developed interactive KPI dashboards reducing redundant issues by 25%
  • Designed Weekly reports/ dashboards on ticket resolution performance which aided the team to optimize resources, resulting in improved performance with 90% of the issues meeting the service level agreement
  • Facilitated daily scrum meetings and monthly review meeting with onsite and offshore team and followed up on action items for continuous improvement

Environment: SQL, SSIS, Microsoft Excel, Tableau

Technical Data Specialist

Confidential

Responsibilities:

  • Responsible for gathering requirements from business analysts and operational analysts, Identified the data sources required for the reports needed to the customers.
  • Used Python programs automated the process of combining the large datasets and Data files and then converting as Teradata tables for Data Analysis.
  • Created an Automated Python Programs to Archive the database tables which large in size and not in use into Mainframes folders.
  • Developed programs with manipulate arrays using libraries like Numpy and Python.
  • Did performance tuning and optimization for increasing the efficiency of the scripts by creating indexes, adding constrains and query optimization
  • Writing SQL scripts for huge data pulls and ad-hoc reports for analysis. Used the Teradata advanced techniques like rank, row number etc.
  • Generated graphs using MS Excel Pivot tables and creating presentations using Power Point.
  • Generated reports using Proc Tab, Proc Report, Data Null, Proc SQL and Macros. Used OLAP functions like sum, count, csum, etc.
  • Communicated with business users and analysts on business requirements. Gathered and documented technical and business Meta data about the data.
  • Created numerous processes and flow charts to meet the business needs and interacted with business users to understand their data needs.
  • Created Set, Multiset, Derived, Volatile, Global Temporary tables to retrieve the data from multiple tables.
  • Experience in writing kornshell scripts for automating the jobs. Automated reports by connecting Teradata from MS Excel using ODBC.
  • Documented scripts, specifications, other processes and preparation of Technical Design Documents.

Environment: Teradata, Teradata utilities (SQL Assistant, BTEQ, Fast Load, Fast Export)

We'd love your feedback!