Data Analyst/data Analytics Resume
Dallas, TX
SUMMARY:
- 10+ years in Information Technology with Expertise in Data modeling for Data Warehouse/Data Mart development, Data Analysis for Online Transaction Processing (OLTP) and Data Warehousing (OLAP)/Business Intelligence (BI) / Data Visualization applications such as SAP Business Objects and Tableau.
- Experience (2 Years) in Analytics - With expertise in Predictive and Prescriptive analytics and data mining techniques with emphasis on Machine Learning algorithms (Supervised and Unsupervised learning) using SAS and Python programming.
- Involved in various projects related to Data Modeling, System/Data Analysis, Design and Development for both OLTP and Data warehousing environments, Data Extraction, Data Cleansing, Data Profiling, Data Mining, Data Consolidation, and Data Quality for various business data feeds.
- Worked extensively on Erwin and ER Studio in several projects in both OLAP and OLTP applications for Data Modeling.
- Expertise in using statistical languages (SAS, Python) to Clean, Integrate, Transform, Reduce, Analyze and Interpret large sets of Data. Experienced in Python's modules numpy, beautiful soup, matplotlib, pandas etc for data preprocessing, web scraping, visualization and for machine learning.
- Exposure in setting up Hadoop Cluster. Configuring HDFS, Hive, Nifi, Solr on Ambari Framework and create dashboard in Zeppelin.
- Facilitated data requirement meetings with business and technical stakeholders and resolved conflicts to drive decisions.
- Practical understanding of the Data modeling (Dimensional & Relational) concepts like Star-Schema Modeling, Snowflake Schema Modeling, Normalization/De-normalization, Fact and Dimension tables, Implemented Slowly Changing Dimensions - Type I & II in Dimension tables as per the requirements.
- Well versed with various types Software Development methodologies- Water fall, Agile, RUP, Iterative and extreme Programming.
- Extensive working experience in RDBMS technologies like Oracle 8.x/9.x/10g/11i, DB 2 8.0/7.0/6.0 , MS SQL Server 2008/2005/2000/7.0/6.5 , Dbase, Excel and MYSQL. SQL Developer, TOAD, SQL Plus, Win SQL, Quest Central for DB2. Good at working with SQL Assistant 7.1 in Teradata environment.
- Ability to develop complicated SQL script for Data validation testing by running SQL script, procedures.
- Established and maintained comprehensive data model documentation including detailed descriptions of business entities, attributes, and data relationships.
- Developed mapping spreadsheets for (ETL) team with source to target data mapping with physical naming standards, datatypes, volumetric, domain definitions, and corporate meta-data definitions.
- Excellent written and oral communication skills and a team-player with a results-oriented attitude.
TECHNICAL SKILLS:
ETL Tools: Informatica Power Center 9,8.5.1 (Designer, Workflow Manager, Workflow Monitor,) Abinitio GDE
OLAP Tools: SAP Business Objects 2010/2008/XIR2, Microstrategy 8.0, IBM Cognos v9
Data Modeling: Erwin r7.1/7.2, ER Studio V8.0.1 and Oracle Designer
Operating Systems: Windows NT/2000/2003/XP, UNIX, Solaris and MSDOS
Databases: Microsoft SQL Server 2000/2005/2008 , DB2, Oracle 10g/9i, MS Access, Teradata
Languages: SQL, T SQL, Base SAS and SAS/SQL, HTML,R, Python
PROFESSIONAL EXPERIENCE:
Confidential, Dallas, TX
Data Analyst/Data Analytics
Responsibilities:
- Exploring and examining the mortgage banking data coming from multiple disparate sources using BI tools such as Business Objects and using Tableau. This helps provide a competitive advantage or address a pressing business problem.
- Performing Data Preprocessing using Python/SAS based on the nature of the source system.
- Performing statistical analysis, data mining and retrieval processes on a large amount of data to identify trends related to Defaulting prediction model, figures and other relevant information using SAS and python.
- Used forward engineering to create a Physical Data Model with DDL that best suits the requirements from the Logical Data Model for Text Analytics from the Existing ICDW RMI data structures.
- Perform Text Analytics to analyze e-mails, surveys and written correspondence to identify customer complaints and metrics associated with it using Python.
- Perform predictive analytics on Mortgage Risk, Loan Defaulting using MB-ICDW RMI data key attributes such as LoanStatus, LoantoValue, TermofMortgage, BorrowerOccupation, GDPGrowth, LoanOriginationDate etc.
- Worked with Database Administrators, Business Analysts and Content Developers to conduct design reviews and validate the developed models, Identified, formulated and documented detailed business rules and Use Cases based on requirements analysis
- Performed due-diligence with the existing data in data warehouse, Generated ad-hoc SQL queries using joins, database connections and transformation rules to compare and evaluate the prediction models.
Confidential, Dallas, TX
Data Modeler/Data Analyst
Responsibilities:
- Studied in-house requirements for the Data warehouse to be developed, Conducted one-on-one sessions with business users to gather data warehouse requirements
- Analyzed database requirements in detail with the project stakeholders by conducting Joint Requirements Development sessions
- Developed a Conceptual model using Erwin based on requirements analysis, Developed normalized Logical and Physical database models to design OLTP system for insurance applications
- Created dimensional model for the reporting system by identifying required dimensions and facts using Erwin r7.1
- Used forward engineering to create a Physical Data Model with DDL that best suits the requirements from the Logical Data Model
- Worked with Database Administrators, Business Analysts and Content Developers to conduct design reviews and validate the developed models, Identified, formulated and documented detailed business rules and Use Cases based on requirements analysis
- Responsible for defining the naming standards for data warehouse, Generated ad-hoc SQL queries using joins, database connections and transformation rules to fetch data from legacy DB2 and SQL Server database systems
- Used Erwin for reverse engineering to connect to existing database and ODS to create graphical representation in the form of Entity Relationships and elicit more information
- Used Informatica Designer, Workflow Manager and Repository Manager to create source and target definition, design mappings, create repositories and establish users, groups and their privileges
- Extracted data from the databases (Oracle and SQL Server, DB2, FLAT FILES) using Informatica to load it into a single data warehouse repository.
- Integrated the work tasks with relevant teams for smooth transition from testing to implementation
Confidential, Dallas, TX
Data Modeler/Data Analytics
Responsibilities:
- Data Munging which facilitates dealing with imperfections in data; mapping data from a raw form into a format that allows for more convenient consumption.
- Solicit, ratify, and document stakeholder requirements specific to data source identification, data quality measurements, data duplication, master record creation and data changes for data domains.
- Profile and analyze customer data against information quality expectations or specifications, looking for patterns of conformance or deviation. Quantify and quality gaps and potential impact.
- Reverse engineered the reports and identified the Data Elements (in the source systems), Dimensions, Facts and Measures required for new enhancements of reports.
- Unwinding the existing logic in SQL Server/Oracle to fit in to new ICDW model and create STT (Source to Target) documents.
- Identify ways of building new rules, insights, metrics, and data improvement approaches. Develop data quality improvement plans that leverage automation as well as other data management best practices.
- Prepare a Base Query and the transformation logic for the respective target Integration Tables/Semantic views
- Developing Source to Target Mapping for the incoming data, certifying the quality of data in tables and files and also in getting the reporting requirement from business to in corporate it to SQL logic
- Creating databases, users, tables, triggers, macros, views, stored procedures, functions, Packages, joins and hash indexes in Teradata database
- Performance analysis, monitoring and SQL query tuning using EXPLAIN PLAN, Collect Statistics, Hints and SQL Trace both in Teradata as well as Oracle.
- Help specify acceptance criteria for test data and ensure test data sampling meetings testing requirements
Confidential, Dallas, TX
Data Analyst
Responsibilities:
- Worked on Requirement Analysis, Data Analysis and Gap Analysis of various source systems sitting and coming from multi systems. Responsible for BI Data Quality.
- Conducted JAD sessions to allow different stakeholders such as editorials, designers, etc.,
- Performed Data Validation with Data profiling, Involved in Data Extraction from Teradata and Flat Files using sql assistant.
- Use SQL, PL/SQL to validate the Data going in to the Data Ware House, Creating complex data analysis queries to troubleshoot issues reported by users
- Evaluates data mining request requirements and help develop the queries for the requests.
- Conducted UAT (User Acceptance Testing) for multiple iterations by writing Test Cases and signed off the same after approval
- Involved in Designing Star Schema, Creating Fact tables, Dimension tables and defining the relationship between them.
- Verified and maintained Data Quality, Integrity, data completeness, ETL rules, business logic.
- Extensively used SQL queries to check storage and accuracy of data in database tables, Used SQL for Querying the Oracle database.
- Worked with business team to test the reports developed in Business Objects.
Confidential
Market Data Analytics (Consultant)
Responsibilities:
- Driving in Algo Remediation Program .i.e., addressing then existing issues such as lack of insight of core algo behavior within the team, missing market specific scenarios in test scope etc by exploring more into each of the algorithm strategies.
- Market Analysis of stocks and its behavior to analyze the performance of algorithms such VWAP, TWAP, POV, PEG, ARID, AQUA etc.
- Creating user specs of new algorithmic models and also helped in successful rollout of these to production without any defects.
- Creating intranet portal highlighting implementation process of each of the algos (on existing behavior/scope/ priority), thereby making it more transparent.
Confidential
Data Analyst
Responsibilities:
- Involved in the Application Study, Creation of Business cases and Execution by Dry run from the Client location during the initial phases of the Trading system Rollout for India Market.
- Planning the Execution cycle based on the Exchange Trading hours and coordinating the same with various Team stockholders involved.
- Deliverable handling by Reviewing Scope, Planning Environment Requirement, Escalating any Test constraints, Resource Allocation and Tracking Schedule slippage if any.