We provide IT Staff Augmentation Services!

Sr Data Analyst Resume

4.00/5 (Submit Your Rating)

New, YorK

SUMMARY:

  • Excellent Software Development Life Cycle (SDLC) with good working knowledge of testing methodologies, disciplines, tasks, resources and scheduling
  • Excellent knowledge in Data Analysis, Data Validation, Data Cleansing, Data Veri cation and identifying data mismatch.
  • Professional experience involving project development, implementation, deployment, and maintenance using Big Data technologies in designing and implementing complete end - to-end Hadoop based data analytical solutions using HDFS, MapReduce, Spark, Scala, Yarn, Kafka, PIG, HIVE, Sqoop, Flume, Oozie, Impala, HBase.
  • Strong working experience in planning and carrying out of Teradata system extraction using Informatica, Loading Process and Data warehousing, Large-scale Database Management and Reengineering.
  • Highly experienced in creating complex Informatica mappings and workflows working with major transformations.
  • Worked on NoSQL databases including HBase and Mongo DB.
  • Experience in Implementing and building CI/CD pipelines with Jenkins and AWS.
  • Experience in using PL/SQL to write Stored Procedures, Functions and Triggers.
  • Having proficient experience in various Big Data technologies like Hadoop, Apache NiFi, Hive Query Language, HBase NoSQL database, Sqoop, Spark, Scala, OOZIE and Pig. Oracle Database and Unix shell Scripting technologies.
  • Experience developing Pig Latin and HiveQL scripts for Data Analysis and ETL purposes and extended the default functionality by writing User Defined Functions (UDFs), User Defined Aggregate Function (UDAFs) for custom data specific processing.
  • Used bunch of steps in Pentaho transformations including Row Normalizer, Row Demoralizer, Database Lookup, Database Join, Calculator, Add Sequence, Add Constants and various types of inputs and outputs for various data sources including Tables, Access, Text File, Excel and CSV.
  • Excellent knowledge on creating reports on Power BI, Tableau, SAP Business Objects, Web reports for multiple data providers
  • Experienced working with Excel Pivot and VBA macros for various business scenarios.
  • Strong experience in Data Analysis, Data Migration, Data Cleansing, Transformation, Integration, Data Import, and Data Export through the use of multiple ETL tools such as Ab Initio and Informatica PowerCenter Experience in testing and writing SQL and PL/SQL statements - Stored Procedures, Functions, Triggers and packages.
  • Experience in automating and scheduling the Informatica jobs using UNIX shell scripting con guring Korn-jobs for Informatica sessions.

PROFESSIONAL EXPERIENCE:

Confidential, New York

Sr Data Analyst

Responsibilities:

  • Analysis of functional and non-functional categorized data elements for data pro ling and mapping from source to target data environment. Developed working documents to support ndings and assign speci c tasks
  • Analysis of functional and non-functional categorized data elements fordata profilingand mapping from source to target data environment. Developed working documents to support findings and assign specific tasks.
  • Implement Spark Kafka streaming to pick up the data from Kafka and send to Spark pipeline.
  • Experience in working with different join patterns and implemented both Map and Reduce Side Joins.
  • Wrote Flume configuration files for importing streaming log data into HBase with Flume.
  • Developed the batch scripts to fetch the data from AWS S3 storage and do required transformations in Scala using Spark framework.
  • Implemented Spark using Scala and SparkSQL for faster testing and processing of data.
  • Monitoring the Hive Meta store and the cluster nodes with the help of Hue.
  • Created AWS EC2 instances and used JIT Servers. and configured pig, written Pig Latin scripts to convert the data from Text file to Avro format.
  • Created Partitioned Hive tables and worked on them using HiveQL.
  • Involved in data ingestion into HDFS using Sqoop for full load and Flume for incremental load on variety of sources like web server, RDBMS and Data API’s.
  • Built the infrastructure required for optimal extraction, transformation, and loading of data from a wide variety of data sources using SQL and ‘big data’ technologies like Hadoop Hive, Azure Data Lake storage
  • Experience in writing shell scripting for various ETL needs.
  • Deep knowledge of RDBMS (SQL Server, MySQL, DB2 etc) and NoSQL databases such as MongoDB, DynamoDB and Cassandra.
  • Experience in creating ETL transformations and jobs using Pentaho Kettle Spoon designer and Pentaho Data Integration Designer and scheduling them on Pentaho BI Server.
  • Used bunch of steps in Pentaho transformations including Row Normalizer, Row Demoralizer, Database Lookup, Database Join, Calculator, Add Sequence, Add Constants and various types of inputs and outputs for various data sources including Tables, Access, Text File, Excel and CSV.
  • Designed, developed, tested, and maintained Tableau reports and dashboards based on user requirements.
  • Created action filters, parameters and calculated sets for preparing dashboards and worksheets. with Drill Down and Drop-down menu option and Parameterized using Tableau.
  • Extensively involved in Developing Data Management strategies for data lineage/metadata, data control data defect/issue.
  • Involve in data analysis, data profiling, metadata management, data governance in an enterprise environment.
  • Involved in designing and building data quality implementation, monitoring and reporting.
  • Created an analytical report that drives ongoing monitoring of resource allocations and utilization.
  • Meeting with user groups to analyze requirements and proposed changes in design and specifications. Performed Detailed Data Analysis DDA, Data Quality Analysis DQA and Data Profiling on source data.
  • Worked on claims data and extracted data from various sources such as at les, Oracle and Mainframes.
  • Worked with data investigation, discovery and mapping tools to scan every single data record from many sources.

Confidential, Minneapolis

Sr Data Analyst

Responsibilities:

  • Working with the Business Analyst and DBA, conducting team meetings and JAD Sessions for technical requirements gathering, business analysis, testing and project coordination.
  • Analyze and evaluate existing system for assessing how it supported functions from the Use Case diagrams and UML diagrams.
  • Developed Logical data model using ERwin 7.2 and created physical data models using forward engineering.
  • Implementation of full lifecycle inData warehousesand Business Data marts withStar Schemas, Snowflake Schemas,SCD&Dimensional Modelling.
  • Strong understanding in the principles of Data ware housing using Fact Tables, Dimension Tables, star schema modelling and snowflake schema modelling, slowly changing dimensions, foreign key concepts, referential integrity.
  • Worked on Provider, Payor and Patient Data pulling from EPIC and Cerner Applications
  • Performed reverse engineering for a wide variety of relational DBMS, including Microsoft Access, Oracle, and Teradata, to connect to existing database and create graphical representation (E-R diagram) using Erwin 7.2.
  • Used Teradata utilities such as Fast Export, MLOAD for handling various tasks.
  • Participate in updating the dimensional model and identify the Facts and Dimensions.
  • Employed process and data analysis to model a Customer Information Business System.
  • Produced functional decomposition diagrams and defined logical data model.
  • Coordinated with ETL team for analysis and reporting needs.
  • Developed data migration strategies and impact analysis for database.
  • Design, plan and program data conversion procedures from old to new databases.
  • Responsible for data lineage, maintaining data dictionary, naming standards and data quality.
  • Model a database to store customer demographic and other data.
  • Work with Business analyst to design weekly reports using a combination of Cognos Impromptu and Power play multidimensional, to deliver reports to CEL and Marketing groups.
  • Design complex Dashboards like Executive, KPI and Strategic by using Tableau.

Confidential, Plano, TX

Data Analyst

Responsibilities:

  • Worked on claims data and extracted data from various sources such as at les, Oracle and Mainframes.
  • Identified the business functions and processes and prepare system scope and objectives based on user needs and industry regulations.
  • Defined terms, conducted stakeholder analysis, elicited business needs, conducted business process modeling work, facilitated work sessions (JAD), elicited requirements, wrote requirements and use cases, analyzed requirements, validate & prioritize requirements, trace requirements to related project documentation (process models, designs, test scenarios & scripts).
  • Collaborated with the Reporting Team to design Monthly Summary Level Cubes to support the further aggregated level of detailed reports.
  • Worked on the Snow-flaking the Dimensions to remove redundancy
  • Designed and implemented the RestAPI for the UI screen through the mobile, user can look up the customer information by taking zip code as the perimeter.
  • Designed a STAR schema for the detailed data marts and Plan data marts involving confirmed dimensions.
  • Created and maintained the Data Model repository as per company standards.
  • Architect & implement medium to large scale BI solutions on Azure using Azure Data Platform services (Azure Data Lake, Data Factory, Data Lake Analytics, Stream Analytics, Azure SQL DW, HDInsight/Databricks, NoSQL DB).
  • Involved in design and development of standard and ad-hoc reporting using SQL.
  • Design and developments of the changes and enhancements. The existing submission and refining programs were enhanced to in corporate the new calculations. Involved in overall data mart testing and different team review meetings.

We'd love your feedback!