We provide IT Staff Augmentation Services!

Sr. Data Analyst Resume

Dorchester, MA

SUMMARY:

  • Experienced Data Analyst and Data modeling Professional with specialization in implementing End - To-End Business Intelligence and Data Warehousing solutions with over 9 year of hands on experience.
  • Experience in Architecture, Design and Development of large Enterprise Data Warehouse (EDW) and Data-marts for target user-base consumption.
  • Expert in writing SQL queries and optimizing the queries in Oracle, SQL Server 2008 and Teradata.
  • Excellent Software Development Life Cycle (SDLC) with good working knowledge of testing methodologies, disciplines, tasks, resources and scheduling.
  • Excellent knowledge in Data Analysis, Data Validation, Data Cleansing, Data Verification and identifying data mismatch.
  • Performed data analysis and data profiling using complex SQL on various sources systems including Oracle and Teradata.
  • Excellent experience on Teradata SQL queries, Teradata Indexes, Utilities such as Mload, Tpump, Fast load and Fast Export.
  • Experience in Dimensional Data Modeling, Star/Snowflake schema, FACT & Dimension tables.
  • Strong experience in using Excel and MS Access to dump the data and analyze based on business needs.
  • Excellent knowledge on Perl & UNIX.
  • Expertise lies in Data Modeling, Database design and implementation of Oracle, AWS Redshift databases and Administration, Performance tuning etc.
  • Experience in analyzing data using Hadoop Ecosystem including HDFS, Hive, Spark, Spark Streaming, Elastic Search, Kibana, Kafka, HBase, Zookeeper, PIG, Sqoop, Flume.
  • Experienced working with Excel Pivot and VBA macros for various business scenarios.
  • Strong experience in Data Analysis, Data Migration, Data Cleansing, Transformation, Integration, Data Import, and Data Export .
  • Experience in cloud development architecture on Amazon AWS, EC2, EC3, Elastic Search, Redshift and Basic on Azure.
  • Data Transformation using Pig scripts in AWS EMR, AWS RDS.
  • Knowledge and working experience on big data tools like Hadoop, Azure, AWS Redshift.
  • Experience working with data modeling tools like Erwin, Power Designer and ER Studio.
  • Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems (RDBMS) and from RDBMS to HDFS.
  • Experience in data analysis using Hive, PigLatin, Impala.
  • Well versed in Normalization / De-normalization techniques for optimum performance in relational and dimensional database environments.
  • Good understanding of AWS, big data concepts and Hadoop ecosystem.
  • Experienced in various Teradata utilities like Fastload, Multiload, BTEQ, and Teradata SQL Assistant.
  • Expert in writing SQL queries and optimizing the queries in Oracle, SQL Server 2008 and Teradata.
  • Develop and manage SQL, Python and R code bases fordatacleansing anddataanalysis using Git version control
  • Excellent Software Development Life Cycle (SDLC) with good working knowledge of testing methodologies, disciplines, tasks, resources and scheduling.
  • Excellent knowledge in Data Analysis, Data Validation, Data Cleansing, Data Verification and identifying data mismatch.
  • Extensive ETL testing experience using Informatica 8.6.1/8.1 (Power Center/ Power Mart) (Designer, Workflow Manager, Workflow Monitor and Server Manager)
  • Have good exposure on working in offshore/onsite model with ability to understand and/or create functional requirements working with client and also have Good experience in requirement analysis and generating test artifacts from requirements docs.
  • Excellent in creating various artifacts for projects which include specification documents, data mapping and data analysis documents.
  • An excellent team player& technically strong person who has capability to work with business users, project managers, team leads, architects and peers, thus maintaining healthy environment in the project.

TECHNICAL SKILLS

Analysis and Modeling Tools: Erwin 9.6/9.5, Sybase Power Designer, Oracle Designer, ER/Studio 9.7.

Database Tools: Microsoft SQL Server 2014/2012 Teradata 15/14, Oracle 12c/11g, MS Access, Poster SQL, Netezza.

OLAP Tools: Tableau, SAP BO, SSAS, Business Objects, and Crystal Reports 9.

ETL Tools: SSIS, Pentaho, Informatica Power 9.6, SAP Business Objects XIR3.1/XIR2, Web Intelligence.

Operating System: Windows, Dos, Unix.

Reporting Tools: Business Objects, Crystal Reports.

Tools & Software’s: TOAD, MS Office, BTEQ, Teradata SQL Assistant.

Big Data Technologies: Hadoop, HDFS 2, Hive, Pig, HBase, Sqoop, Flume.

Cloud: AWS, EC2, S3, SQS, Azure.

Other tools: TOAD, SQL PLUS, SQL LOADER, MS Project, MS Visio and MS Office, Have worked on C++, UNIX, PL/SQL etc.

PROFESSIONAL EXPERIENCE

Confidential, Dorchester, MA

Sr. Data Analyst

Responsibilities:

  • Working in Regulatory Compliance IT team where worked as Data Analyst role which involved Data Profiling, Data Modeling, ETL Architecture & Oracle DBA.
  • Responsible for Big data initiatives and engagement including analysis, brainstorming, POC, and architecture.
  • Designed the Logical Data Model using ERWIN 9.64 with the entities and attributes for each subject areas.
  • Developed long term data warehouse roadmap and architectures, designs and builds the data warehouse framework per the roadmap.
  • Working on Amazon Redshift and AWS and architecting a solution to load data create data models and run BI on it.
  • Designed the Data Marts in dimensional data modeling using star and snowflake schemas.
  • Developed Shell, Perl and Python scripts to automate and provide Control flow to Pig scripts.
  • Involved in building database Model, APIs and Views utilizingPython, in order to build an interactive web based solution.
  • Working on logical and physical modeling, and ETL design for manufacturingdatawarehouse applications.
  • Involved in creating Hive tables, and loading and analyzing data using hive queries Developed Hive queries to process the data and generate the data cubes for visualizing Implemented.
  • Implemented Join optimizations in Pig using Skewed and Merge joins for large datasets schema.
  • Designed and developed a Data Lake using Hadoop for processing raw and processed claims via Hive and Informatica.
  • Architected, Designed and developed elastic index, an ingestion framework to index useful data to elastic search and search capability of Common Analytic Framework using elastic search.
  • Developed and implemented different Pig UDFs to write ad-hoc and scheduled reports as required by the Business team.
  • Participated in OLAP model based on Dimension and FACTS for efficient loads of data based on Star Schema structure on levels of reports using multi-dimensional models such as Star Schemas and Snowflake Schema.
  • Design of Big Data platform technology architecture. The scope includes data intake, data staging, data warehousing, and high performance analytics environment.
  • Worked on Amazon Redshift and AWS and architecting a solution to load data, create data models
  • Creating dimensionaldatamodels based on hierarchical sourcedataand implemented on Teradata achieving high performance without special tuning.
  • Involved in loading data from LINUX file system to HDFS Importing and exporting data into HDFS and Hive using Sqoop Implemented Partitioning, Dynamic Partitions, Buckets in Hive.
  • Working on definingdataarchitecture fordatawarehouses,Datamarts and business applications.
  • Designed and developed architecture fordataservices ecosystem spanning Relational, NoSQL, and BigDatatechnologies.
  • Installing and configuring the a 3-node Cluster in AWS EC2 Linux Servers.
  • Specifies overall Data Architecture for all areas and domains of the enterprise, including Data Acquisition, ODS, MDM, Data Warehouse, Data Provisioning, ETL, and BI.
  • Developed Data Mapping, Data Governance, and Transformation and cleansing rules for the Master Data Management Architecture involving OLTP, ODS.
  • Worked on batch processing of data sources using Apache Spark, Elastic search.
  • Working ondataprofiling and analysis to create test cases for new Architecture evaluation.
  • Involved in Normalization / De normalization techniques for optimum performance in relational and dimensional database environments.
  • Implementing multiple large scale DW/BI projects involvingDataIntegration, Migration, Conversion, Upgrade and Reporting (BI) projects.
  • Data Profiling, Mapping and Integration from multiple sources to AWS S3.
  • Performance tuning and stress-testing of NoSQL database environments in order to ensure acceptable database performance in production mode.
  • Involved in writing Liquid base scripts and generating SQL's.
  • Implemented strong referential integrity and auditing by the use of triggers and SQL Scripts.
  • Created and managed database objects (tables, views, indexes, etc.) per application specifications. Implemented database procedures, triggers and SQL scripts for development teams.
  • Designed and developed T-SQL stored procedures to extract, aggregate, transform, and insert data.
  • Created and maintained SQL Server scheduled jobs, executing stored procedures for the purpose of extracting data from DB2 into SQL Server.
  • Developed SQL Stored procedures to query dimension and fact tables in data warehouse.
  • Experience with SQL Server Reporting Services (SSRS) to author, manage, and deliver both paper-based and interactive Web-based reports.
  • Performed Hive programming for applications that were migrated to big data using Hadoop
  • Deployed SSRS reports to Report Manager and created linked reports, snapshots, and subscriptions for the reports and worked on scheduling of the reports.
  • Generated parameterized queries for generating tabular reports using global variables, expressions, functions, and stored procedures using SSRS.
  • Used SSRS to create reports, customized Reports, on-demand reports, ad-hoc reports and involved in analyzing multi-dimensional reports in SSRS.
  • Building dashboards using SSRS 2008 Build, test, and execute one-off procedures as needed to support business commitments. Answer business user questions. Investigate and resolve issues or problems that arise. On call for production troubles.
  • Generate comprehensive analytical reports by running SQL queries against current databases to conduct data analysis pertaining to various loan products.
  • Created SQL codes from data models and interacted with DBA's to create development, testing and production database.
  • Extensive knowledge in Data loading using PL/ SQL Scripts and SQL Server Integration Services (SSIS).
  • Work in team using ETL tool Informatica to populate the database, data transformation from the old database to the new database using Oracle and SQL Server.
  • Focused on architecting NoSQL databases like Mongo, Cassandra and Cache database.
  • Perform routine management operations, including configuration and performance analysis for mongodb. Diagnosing Performance Issues for mongodb.
  • Creating MongoDB data set backups using system-level file snapshot tool, such as LVM or native storage appliance tools.
  • Point in time Backup and recovery in MongoDB using MMS. Data modeling for data from RDBMS to and MongoDB for optimal reads and writes.
  • Designed and developed architecture fordataservices ecosystem spanning Relational, NoSQL, and BigDatatechnologies.
  • Involved in designing Logical and Physical data models for different database applications using the Erwin.
  • Data modeling, Design, implement, and deploy high-performance, custom applications at scale on Hadoop.
  • Executed Hive queries on Parquet tables stored in Hive to perform data analysis to meet the business requirements.
  • Developed Map Reduce programs to cleanse the data in HDFS obtained from heterogeneous data sources to make it suitable for ingestion into Hive.
  • Implemented Data Integrity and Data Quality checks in Hadoop using Hive and Linux scripts.
  • Applied data analysis, data mining and data engineering to present data clearly.
  • Reverse engineered some of the databases using Erwin.
  • Proficiency in SQL across a number of dialects (we commonly write MySQL, PostgreSQL, Redshift, SQL Server, and Oracle).
  • Routinely deal in with large internal and vendor data and perform performance tuning, query optimizations and production support forSAS, Oracle 12c.

Environment: DB2, CA Erwin 9.6, Oracle 12c, MS-Office, SQL Architect, TOAD Benchmark Factory, SQL Loader, PL/SQL, SharePoint, ERwin r9.64, Talend, MS-Office, Redshift, SQL Server 2008/2012, Hive, Pig, Hadoop, AWS.

Confidential, Chicago, IL

Sr. Data Analyst / Modeler

Responsibilities:

  • Performedas a Data Analysis, Data Modeling, Data Migration and data profiling using complex SQL on various sources systems including Oracle andTeradata.
  • Managed Logical and Physical Data Models in ER Studio Repository based on the different subject area requests for integrated model.DevelopedDataMapping,DataGovernance, and transformation and cleansing rules involving OLTP, ODS.
  • Enforced referential integrity in the OLTPdatamodel for consistent relationship between tables and efficient database design.
  • Full life cycle ofDataLake,Data Warehouse with Bigdatatechnologies like Spark, Hadoop.
  • Establish and develop new database maintenance scripts -Automate Netezza database management and monitoring.
  • Fulfilled requirements of multiple consumers of the data by taking into account the usage and statistical models design into consideration for Neo4j, SAS, Azure.
  • Designed Star and Snowflake Data Models for Enterprise Data Warehouse using ER Studio.
  • Utilized Informatica toolset (Informatica Data Explorer, and Informatica Data Quality) to analyze legacy data for data profiling.
  • Extracted Mega Data from AWS, and Elastic Search engine using Sql Queries to create reports.
  • Performed database health checks and tuned the databases using Teradata Manager.
  • Experience using MapReduce, and "Big data" work on Hadoop and other NOSQL platforms.
  • Developed, managed and validated existingdatamodels including logical and physical models of theData Warehouse and source systems utilizing a 3NF model.
  • Used Star schema and Snowflake Schema methodologies in building and designing the LogicalDataModel into Dimensional Models.
  • Worked on Azure and Azure data lake and architecting a solution to load data, create data models.
  • Implemented logical and physical relational database and maintained Database Objects in thedatamodel using ER Studio.
  • Involved in debugging and Tuning the PL/SQL code, tuning queries, optimization for the Oracle, and DB2 database.
  • Worked with data investigation, discovery and mapping tools to scan every single data record from many sources.
  • Good understanding and hands on experience in setting up and maintainingNoSQLDatabases like Cassandra and HBase.
  • Selecting the appropriate Azure service based on data, compute, database, or security requirements
  • Developed automateddatapipelines from various externaldatasources (web pages, API etc) to internalData Warehouse (SQL server,AWS), then export to reporting tools like Datorama by Python.
  • Developed several behavioral reports and data points creating complex SQL queries and stored procedures using SSRS and Excel.
  • Generated periodic reports based on the statistical analysis of the data using SQL Server Reporting Services (SSRS).
  • Conducted numerous POCs (Proof of Concepts) to efficiently import large data sets into the database from Azure.
  • Generated reports using Global Variables, Expressions and Functions using SSRS.
  • Developed different kind of reports such as Drill down, Drill through, Sub Reports, Charts, Matrix reports, Parameterized reports and Linked reports using SSRS.
  • Created dimensional model for reporting system by identifying required dimensions and facts using Erwin r8.0.
  • Extensively used ERwin for developing data model using star schema methodologies
  • Collaborated with other data modeling team members to ensure design consistency and integrity.
  • Involved in Planning, Defining and Designing data base using Erwin on business requirement and provided documentation.
  • Defined and deployed monitoring, metrics, and logging systems on Azure.
  • Validated the data of reports by writing SQL queries in PL/SQL Developer against ODS.
  • Involved in user training sessions and assisting in UAT (User Acceptance Testing).
  • Strong ability in developing advanced ANSI SQL queries to extract, manipulate, and/or calculate information to fulfill data and reporting requirements including identifying the tables and columns from which data is extracted.
  • Definemetadatabusiness-level (logical) terms through interactions with project teams, business subject matter experts, and data analysis.

Environment: PL/SQL, Business Objects XIR3, ER Studio, NoSQL, ETL Tools Informatica 8.6, Oracle 12c/11g, AmazonRedshif, Teradata V2R14, Teradata SQL Assistant 12.0, ETL Tools, Azure.

Confidential, Bronx, NY

Sr. Data Analyst

Responsibilities:

  • Designed and Developed logical & physical data models and Meta Data to support the requirements using Erwin
  • Designed the ER diagrams, logical model (relationship, cardinality, attributes, and, candidate keys) and physical database (capacity planning, object creation and aggregation strategies) for Oracle and Teradata as per business requirements using Erwin
  • Designed 3rd normal form target data model and mapped to logical model.
  • Involved in extensive DATA validation using ANSI SQL queries and back-end testing
  • Generated DDL statements for the creation of new ER/studio objects like table, views, indexes, packages and stored procedures.
  • Design MOLAP/ROLAP cubes on Teradata Database using SSAS.
  • Used SQL for Querying the database in UNIX environment
  • Creation of BTEQ, Fast export, Multi Load, TPump, Fast load scripts for extracting data from various production systems .
  • Worked RDS for implementing models and data on RDS.
  • Developed mapping spreadsheets for (ETL) team with source to target data mapping with physical naming standards, data types, volumetric, domain definitions, and corporate meta-data definitions.
  • Experienced in Using CA Erwin Data Modeler (Erwin) for Data Modeling (data requirements analysis, database design etc.) of custom developed information systems, including databases of transactional systems and data marts.
  • Designing Star schema and Snow Flake Schema on Dimensions and Fact Tables
  • Expertise in Informatica, DB2, Microstrategy and UNIX Shell scripting
  • Worked with Data Vault Methodology Developed normalized Logical and Physical database models.
  • Transformed Logical Data Model to Physical Data Model ensuring the Primary Key and Foreign key relationships in PDM, Consistency of definitions of Data Attributes and Primary Index considerations.
  • Wrote and running SQL, BI and other reports, analyzing data, creating metrics/dashboards/pivots/etc.
  • Gather and analyze business data requirements and model these needs. In doing so, work closely with the users of the information, the application developers and architects, to ensure the information models are capable of meeting their needs.
  • Working along with ETL team for documentation of transformation rules for data migration from OLTP to warehouse for purpose of reporting.
  • Created views and extracted data from Teradata base tables and uploaded data to oracle staging server from Teradata tables, using fast export concept.

Environment:,, SQL Server, Erwin9.1, Oracle, Informatica, RDS, Big Data, JDBC, NOSQL, Spark, Scala, Star schema, Snow Flake Schema, Python, MySQL, PostgreSQL .

Confidential, NYC, NY

Data Analyst / Modeler

Responsibilities:

  • Analyzed the physicaldatamodel to understand the relationship between existing tables. Cleansed the unwanted tables and columns as per the requirements as part of the duty being aDataAnalyst.
  • Developed data Mart for the base data in Star Schema, Snow-Flake Schema involved in developing the data warehouse for the database.
  • Worked on Unit Testing for three reports and created SQL Test Scripts for each report as required
  • Extensively used ER Studio as the main tool for modeling along with Visio
  • Established and maintained comprehensive data model documentation including detailed descriptions of business entities, attributes, and data relationships.
  • Involved in understanding and creating Logical and Physical Data model using ERstudio Tool.
  • Created dynamic SSRS reports with the help of parameters which changes according to the user input.
  • Created various SSRS dashboard reports for commission, sales and profit analysis.
  • Designed Star and Snowflake Data Models for Enterprise Data Warehouse using ER Studio.
  • Worked on Metadata Repository (MRM) for maintaining the definitions and mapping rules up to mark .
  • TrainedSpotfiretool and gave guidance in creatingSpotfireVisualizations to couple of colleagues
  • Developed Contracting Business Process Model Workflows (current / future state) usingBizagi Process Modeler software.
  • Created a Data Mapping document after each assignment and wrote the transformation rules for each field as applicable
  • Worked on Unit Testing for three reports and created SQL Test Scripts for each report as required
  • Configured & developed the triggers, workflows, validation rules & having hands on the deployment process from one sandbox to other.
  • Modified cube and dimensions, deployed KPIs on SSRS and created different metrics, added new measure groups and dimensions.
  • Managed Logical and Physical Data Models in ER Studio Repository based on the different subject area requests for integrated model.
  • Created automatic field updates via workflows and triggers to satisfy internal compliance requirement of stamping certain data on a call during submission.
  • Established and maintained comprehensive data model documentation including detailed descriptions of business entities, attributes, and data relationships.
  • Worked on Metadata Repository (MRM) for maintaining the definitions and mapping rules up to mark.
  • Developed data Mart for the base data in Star Schema, Snow-Flake Schema involved in developing the data warehouse for the database.
  • Developed enhancements toMongo DBarchitecture to improve performance and scalability.
  • Forward Engineering the Data models, Reverse Engineering on the existing Data Models and Updates the Data models.
  • Performeddatacleaning anddatamanipulation activities using NZSQL utility.
  • Analyzed and understood the architectural design of the project in a step by step process along with the data flow.
  • Created DDL scripts for implementing Data Modeling changes. Created ERWIN reports in HTML, RTF format depending upon the requirement, Published Data model in model mart, created naming convention files, co-coordinated with DBAs' to apply the data model changes.

Environment: Oracle SQL Developer, OracleDataModeler, Teradata14, SSIS, Business Objects, SQL Server 2008, ER/Studio Windows, MS Excel.

Confidential

Data Analyst / Modeler

Responsibilities:

  • Attended and participated in information and requirements gathering sessions
  • Translated business requirements into working logical and physical data models for Data Warehouse, Data marts and OLAP applications.
  • Performed extensive Data Analysis and Data Validation on Teradata.
  • Designed Star and Snowflake Data Models for Enterprise Data Warehouse using ERWIN
  • Created and maintained Logical Data Model (LDM) for the project.Includes documentation of all entities, attributes, data relationships, primary and foreign key structures, allowed values, codes, business rules, glossary terms, etc.
  • Experienced in Designed and Developed logical & physicaldatamodels and MetaDatato support the requirements using ERWIN.
  • Involved using ETL tool Informatica to populate the database,datatransformation from the old database to the new database using Oracle.
  • Involved in the creation, maintenance ofData Warehouse and repositories containing Metadata.
  • Wrote and executed unit, system, integration and UAT scripts in a Data Warehouse projects.
  • Wrote and executed SQL queries to verify that data has been moved from transactional system to DSS, Data Warehouse, data mart reporting system in accordance with requirements.
  • Excellent experience and knowledge on Data Warehouse concepts and dimensional data modelling using Ralph Kimball methodology.
  • Developed separate test cases for ETL process (Inbound & Outbound) and reporting.

Environment: Oracle, MS Visio, PL-SQL, Microsoft SQL Server 2000, Rational Rose, Datawarehouse, OLTP, OLAP, ERWIN, Informatica 9.x, Windows, SQL, PL/SQL, SQL Server, TalendDataQuality, Talend Integration Suite 4x, Oracle 9i/10g, Flat Files, Windows, SVN.

Hire Now