Data Analyst Resume
Wellesley Hills, MA
SUMMARY
- Experience as a Data Analyst with solid understanding of Data Modeling, Evaluating Data Sources and strong understanding of Data Warehouse/Data Mart Design, ETL, BI, OLAP, Client/Server applications to implement management and staff's business requirements into the software application in Banking and Finance Industry.
- Excellent knowledge and experience in data warehouse development life cycle, SDLC, repository management and administration, dimensional modeling, implementation of Star and Snowflake schemas.
- Experience with Azure Big Data Systems like Azure Data lake Store, Azure Data Analytics, HD Insights, Azure Databricks, Azure Machine Learning, U - SQL scripts, Spark, Python.
- Proficient in report writing and data visualization tools.
- Deep understanding of Investment Management, Asset Management, Prime Brokerage Services, Money and Capital Market Instruments.
- Extensive experience in Strategic development of a Data Warehouse and in Performing Data Analysis and Data Mapping from an Operational Data Store to an Enterprise Data Warehouse.
- Experience in Python with proven expertise in using new tools and technical developments
- Experience in Apache Spark cluster and streams processing using Spark Streaming
- Worked on several python packages like NumPy, scipy, pandas, pytables etc.
- Excellent knowledge in Data Analysis, Data Validation, Data Cleansing, Data Verification and identifying data mismatch.
- Sound understanding of Big Data Analytics and its technologies, Digital and IoT concepts, AWS Cloud. SMAC (Social, Mobility, Analytics and Cloud)
- Expert in Data Modeling, Data Analysis, Data Visualization and Modern Data Warehouse concepts.
- Designed Various Reports/dashboards to provide insights and data visualization using BI/ETL tools like Business Objects, Tableau, and Pentaho.
- Knowledge in the ETL (Extract, Transform and Load) of data into a data ware house/date mart and Business Intelligence (BI) tools like Business Objects Modules (Reporter, Supervisor, Designer, and Web Intelligence).
- Big data management with Power Pivot, working with diverse data sources and building custom interactive dashboards with Tableau to answer business needs.
- Built an automated screenshot comparison tool in Python for web testing.
- Expert Enterprise Data Modeler with a deep understanding of developing transactional enterprise data models that strictly meet normalization rules, as well as enterprise data warehouses using Kimball and Inmon data warehouse methodologies.
TECHNICAL SKILLS
Data analysis: Requirements Gathering, JAD sessions, Process/Production Model analysis, Data Normalization, Cleansing, Profiling, System Design, Data architecture internal Standards development, Metadata and Reports, Source and Target system Analysis
Languages: Python, SQL, T-SQL, PL/SQL, UNIX
Big Data technologies: AWS, HBase 1.2, HDFS, Sqoop 1.4, Spark, Hadoop 3.0, Hive 2.3
MS Office Suite: MS Word, MS PowerPoint, MS Excel, MS Access
Database Systems: SQL Server, Oracle, Teradata, DB2
Operating Systems: Microsoft Windows, Linux, Unix
ETL and Reporting Environment: SQL Server, SSIS, SSRS, Informatica, SAS, Data Stage, QlikView, Tableau
PROFESSIONAL EXPERIENCE
Confidential, Wellesley Hills, MA
Data Analyst
Responsibilities:
- Worked with Data Marts, Data warehousing, Operational Data Store (ODS), OLAP, Star Schema Modeling, Snow-Flake Modeling, Fact and Dimensions Tables.
- Extensively Worked on Star Schema, Snowflake Schema, Data Modeling, Logical and Physical Model, Data Elements, Issue/Question Resolution Logs, and Source to Target Mappings, Interface Matrix and Design elements.
- Worked as Data Analyst to turn data into information, information into insight and insight into business decision.
- Imported/exported large amounts of data from files to Teradata and vice versa.
- Performed numerous data pulling requests using SQL for analysis.
- Extensively involved in the Analysis, design and Modeling. Worked on Snowflake Schema, Data Modeling, Data Elements, Issue/Question Resolution Logs, and Source to Target Mappings, Interface Matrix and Design elements.
- Designed and developed logical and physical data models that utilize concepts such as Star Schema, Snowflake Schema and Slowly Changing Dimensions
- Captured data lineage for all the top-level reports by validating the authorized data sources with system of records and system of original
- Worked on AWS Cloud management and responsible for code build, release and configuration on Amazon EC2.
- Created highly scalable, resilient, and performance architecture using Amazon AWS cloud technologies such as Simple Storage Service (S3), Elastic MapReduce (EMR), Elastic Cloud Compute (EC2), Elastic Container Service (ECS), Lambda, and Elastic Load Balancing (ELB).
- Created SQL-Loader scripts to load legacy data into Oracle staging tables and wrote SQL queries to perform Data Validation and Data Integrity testing.
- Generated reports using the Teradata advanced techniques like rank, row number and etc.
- Involved in Master data analysis, design, Interfaces analysis, Data Analysis, Data Quality, Data Architecture tasks.
- Performed Data Analysis and Data Validation by writing complex SQL queries using Teradata SQL Assistant.
- Involved in mentoring specific projects in application of the new SDLC based on the Agile Unified Process, especially from the project management, requirements and architecture perspectives.
- Worked on Data mapping, logical data modeling, used SQL queries to filter data within the Oracle database tables.
- Revised the Architecture document to reflect Marriott standards for ETL and ELT processes, ODS loading, the Netezza data warehouse environment and extraction via interfaces, reports and querying.
- Assisted in the analysis, design, coding and development of building a new schema (staging tables, views, SQL files).
- Facilitated creation of Taxonomy of the site with mapping of current metadata with new PRISM elements.
- Created Automation scripts using Python.
- Authored various Use Cases and Activity diagrams, Sequence diagrams using Rational Requisite Pro and used UML methodology to define the Data Flow Diagrams (DFD).
- Created tables, views, database links to see the information extracted from SQL files.
- Worked on tables (Like Set, Multiset, Derived, Volatile, Global Temporary), views, using SQL scripts
- Wrote and executed unit, system, integration and UAT scripts in a Data Warehouse projects.
- Responsible for creating Hive tables, loading data and writing hive queries.
- Created Logical/Physical Data models in 3NF in the Warehouse area of Enterprise Data Warehouse.
- Performed Data analysis using Python Pandas.
- Identified/documented data sources and transformation rules required to populate and maintain data warehouse content.
- Built models in spark using python in Databricks.
Confidential, Dorchester, MA
Data Analyst
Responsibilities:
- Involved in Logical & Physical Data Modeling. Database Schema design and modification of Triggers, Scripts, Stored Procedures in Sybase Database Servers.
- Developed monitoring and notification tools using Python.
- Wrote Python routines to log into the websites and fetch data for selected options.
- Worked on data modelling and produced data mapping and data definition documentation.
- Imported the claims data into Python using Pandas libraries and performed various data analysis.
- Trained analysts on how to securely transmit and share documents in the cloud and bank intranet.
- Source Data Analysis, Source to Target data mapping, Data Integration, Define Data Governance requirements.
- Participated in project review meetings and gathering Data Analysis Documents.
- Conducted a budget planning of the project through walkthroughs and meetings involving various leads from Development, QA and Technical Support teams.
- Developed Views and Templates with Python and using Django's view controller and template language, Website interface is created.
- Understood the business process, performed As-Is and was the prime resource for any processes at macro and micro levels.
- Assist other project teams with data analysis, data mining and data profiling as needed.
- Worked on requirements statements, use cases, process flows, site maps, taxonomy/ontology analysis, and wire frames.
- Worked on AWS services such as (EC2, EBS, ASG, RDS, R53)
- Developed AWS Cloud Formation templates to create custom sized VPC, subnets, EC2 instances, ELB, Security Groups.
- Created a high availability and scalable AWS stacks using EC2 auto scaling functionality.
- Developed data conversion programs for membership, claims, and benefit accumulator data - converted thirteen corporate acquisitions.
- Developed data field mappings. Provided programming and support for claims processing functions and auto-adjudication.
- Data mapping, logical data modeling, created class diagrams and ER diagrams and used SQL queries to filter data within the Oracle database.
- Performed several ad-hoc data analyses in Azure Databricks Analysis Platform on KANBAN board.
- Automated daily reporting tasks by creating jobs in Databricks using python and spark.
- Experience in using collections in Python for manipulating and looping through different user defined objects.
- Developed and executed User Acceptance Testing portion of test plan.
- Generated Python Django forms to record data of online users and used PyTest for writing test cases.
- Prepared Logical Data Models that contains set of Entity Relationship Diagrams and Data Flow Diagrams and supporting documents and descriptions of the Relationships between the data elements to analyze and document the Business Data Requirements.
- Prepared test Data sets and performed data testing using the PL/SQL scripts. Also used MS excel for data mining, data cleansing, data mapping, data dictionary and data analysis.
- Mapped process flows; assess as-is processes through user interviews, data collection and analysis, design and evaluation of to-be process solutions.
- Worked on Databricks and Crescendo for Automation
- Authored progress and completion reports, which were then submitted to project management on a weekly basis.
- Participated in writing data mapping documents and performing gap analysis to verify the compatibility of existing system with new business requirements
Confidential, Providence, RI
Data Analyst
Responsibilities:
- Developed Code Using Big Data Software Technologies (Spark, AWS, Databricks) to conduct the analysis of billions of customer records.
- Experience and good knowledge in AWS (Amazon Web Services) services like EC2, S3, Glacier, Elastic Load Balancer (ELB), RDS, SNS, SWF, Cloud watch, Route53 and Lambda.
- Performed AWS Cloud administration managing EC2 instances, S3, SES and SNS services.
- Used Python to place data into JSON files for testing Django Websites.
- Updated and manipulated content and files by using python scripts.
- Responsible for prepare plan and data set for data migration from one legacy system to newly proposed system.
- Resolved user queries on the data present in Data Warehouse, Hadoop and AWS Cloud and help the users in providing the definitions and data usage.
- Involved in Data Analysis, Data Mapping and Data Modeling and Performed data integrity, Data Portability testing by executing SQL statements.
- Used Python scripts to update content in the database and manipulate files.
- Generated Python Django Forms to record data of online users
- Used Python and Django creating graphics, XML processing, data exchange and business logic implementation
- Collaborated with corporate accountants to analyze Financial Statements for Risk Estimation according to GAAP principles.
- Created Use case Diagrams, Activity Diagrams, Sequence Diagrams and ER Diagrams in MS Project.
- Involved in Data Modeling of both Logical Design and Physical design of data warehouse and data marts in Star Schema and Snow Flake Schema methodology.
- Created daily and monthly reports using SQL and UNIX for Business Analysis.
- Utilized ODBC for connectivity to Teradata & MS Excel for automating reports and graphical representation of data to the Business and Operational Analysts.
- Performed extensive data modelling to differentiate between the OLTP and Data Warehouse data models
- Test automation experience for GUI and application interface regression test using Win Runner, QTP and Mercury Quality Center, and scripting (TCL, Python and Perl); also experience on DB/SQL system (Oracle, tec.).
- Requirements Analysis: Gap-Fit each MDM requirement as well as provide high level of estimate for the level of effort required to meet each requirement.
- Designed and Developed the Business Objects Universes which suit the standard, analytical and ad-hoc reporting requirements of the Business Objects users.
- Skilled in using collections in Python for manipulating and looping through different user defined objects.
- Used the DataStage Designer to develop processes for extracting, cleansing, transforms, integrating and loading data into data warehouse database.
- Used Star Schema and Snow flake Schema for data marts / Data Warehouse.
- Gathered requirements and modeled the data warehouse and the underlying transactional database.
- Facilitated Joint Requirement Planning (JRP) sessions with SME's in understanding the Requirements pertaining to Loan Origination to Loan Processing.