We provide IT Staff Augmentation Services!

Data Analyst / Business Analyst Resume

0/5 (Submit Your Rating)

CT

SUMMARY

  • Skilled and experienced data analyst with over 9 years of experience in the field.
  • I have a proven track record of successfully transitioning applications from legacy data sources to advanced and reliable data sets through reengineering efforts.
  • I have extensive experience in utilizing SQL development tools, including
  • Excellent understanding and working experience of industry - standard methodologies like System Development Life Cycle (SDLC) AGILE, and Waterfall Methodologies.
  • I have a strong background in large-scale, highly available distributed database systems and have experience in SQL optimization and performance tuning.
  • I have also led teams in migrating applications from one database platform to another, resulting in cost savings and improved performance
  • Strong expertise in Relational Database systems like Oracle, MS SQL Server, MS Access, DB2 design, and database development using SQL, PL/SQL, SQL PLUS, TOAD, and SQL-LOADER. Highly proficient in writing, testing, and implementing triggers, stored procedures, functions, packages, and Cursors using PL/SQL.
  • Developed Python scripts to parse the Flat Files, CSV, XML, JSON files and extract the data from various sources and load the data into the data warehouse.
  • Developed automated scripts to do the migration using UNIX shell scripting, Python, Oracle/TD SQL, TD Macros, and Procedures.
  • Experience in Data Integration and Data Warehousing using various ETL tools and Informatics
  • Extensive experience in all stages of SDLC (Agile, Waterfall), writing Technical Design documents, Development, Testing, and Implementation of Enterprise level Data mart and Data warehouses.
  • Experience in Designing Business Intelligence Solutions with Microsoft SQL Server and using MS SQL Server Integration Services (SSIS), MS SQL Server Reporting Services (SSRS), and SQL Server Analysis Services (SSAS).

TECHNICAL SKILLS

Databases: Expertise in Oracle, and PostgreSQL with experience in database design, data modeling, and installation.

Query Skills: Strong skills in SQL, and PLSQL with experience in query optimization, performance tuning, creating functions, procedures, and partitions.

Tools: Data bricks, JIRA

ETL: Proficient in Informatica and DBT.

Scripting: Skilled in Linux and Python scripting.

Version Control: Familiar with Git, Stash, and CI/CD for version control.SDLC AGILE, WATERFALL, SCRUM, BRD, BRS, SDLC

PROFESSIONAL EXPERIENCE

Confidential, CT

Data Analyst / Business Analyst

Responsibilities:

  • Extensive experience with large-scale, highly available distributed database systems, including SQL, PL/SQL optimization, and performance tuning
  • Successfully migrated the application from Teradata to Exadata, resulting in cost savings and improved performance
  • Designed and developed Security Framework to provide fine-grained access to objects in AWS S3 using AWS Lambda, and Dynamo DB.
  • Performed end-to-end Architecture & implementation assessment of various AWS services like Amazon EMR, Redshift, and S3.
  • Worked on Amazon AWS concepts like EMR and EC2 web services for fast and efficient processing of Big Data.
  • Working with Informatica Designer to build ETL code, make changes in the existing mappings, adding lookups & logic in other Transformations as per the requirements.
  • Worked on AWS-hosted Data bricks environment and used spark structured streaming to consume the data from Kafka topics in real time and perform merge operations on delta lake tables.
  • Good understanding of Spark Architecture with Data bricks, and Structured Streaming.
  • Hands-on experience in migrating Greenplum to data bricks.
  • Used AWS EMR to transform and move large amounts of data into and out of other AWS data stores and databases, such as Amazon Simple Storage Service (Amazon S3) and Amazon Dynamo DB.
  • Implemented Spark using Scala and utilized Spark SQL heavily for faster development, and processing of data.
  • Planning, scheduling, and implementing Oracle to MS SQL server migrations for AMAT in-house applications and tools.
  • Developed Spark/Scala, Python for regular expression (regex) project in the Hadoop/Hive environment with Linux/Windows for big data resources.
  • Worked on Snowflake for sql development, stored procedures and data analysis, and loading data using DBT models.
  • Worked with PowerShell and UNIX scripts for file transfer, emailing, and other file-related tasks.
  • AWS Elastic Container Service (ECS) was implemented to automate application deployment in the cloud.
  • Experience in using the AWS services Athena, Redshift, and Glue ETL jobs.
  • Experience working with data structures, ETL pipelines, and data processing using DBT
  • Connection with Oracle, MySQL Database is implemented using Hibernate ORM. Configured hibernate, entities using annotations from scratch.
  • Worked on different Python libraries such as Pandas, NumPy, etc.,
  • Collaborated with cross-functional teams to understand business requirements and design data solutions
  • CI/CD using Git, Docker, Jenkins, and version control tools like Stash

Confidential, MI

Data Analyst / Business Analyst

Responsibilities:

  • Automated existing PL SQL packages and re-designed modules to improve performance with SQL/PLSQL optimization and query-tuning techniques
  • Designed and developed Spark workflows that extract data from an AWS S3 bucket and apply transformations to it using Scala and Snowflake.
  • Developed Python AWS server less lambda with concurrent and multi-threading to make the process faster and asynchronously execute the callable.
  • I worked with two different teams, and projects and on-premises application was migrated to AWS.
  • I've used AWS services like EC2 and S3 to process and store small data sets, and I've also worked with the Hadoop cluster on AWS EMR.
  • Monitored containers in AWS EC2 machines using Data Dog API and ingest, and enrich data into the internal cache system.
  • Involved in a project with a team, to migrate Greenplum to data bricks.
  • Design, Development, and Implementation of ETL process to support CDC- Change Data Capture on Data Bricks platform.
  • Responsible for estimating the cluster size, monitoring, and troubleshooting the Spark Data bricks cluster.
  • Converted current AWS architecture and wrote Terraform scripts to automate AWS services such as ELB,
  • Cloud Front distribution, RDS, EC2, database security groups, and S3 bucket.
  • Created multiple Recipes in Glue Data Brew and then used them in various Glue ETL Jobs.
  • Ingestion data through AWS Kinesis Data Stream and Firehose from various sources to S3.
  • Write SQL queries, and create test data in CRM for Informatica cloud mappings unit testing
  • Designed and implemented data pipelines on AWS for data flow from various sources to the Data Lake and data warehouses using AWS Glue, Athena, and S3.
  • Developed Job Scheduler scripts for data migration using UNIX Shell scripting.
  • Used Spark-SQL to load JSON data and create schema RDD and loaded it into Hive tables handled structured data using Spark SQL.
  • Involved in migration projects to migrate data from data warehouses on Oracle/DB2 and migrated those to Teradata.
  • Developed custom ETL solutions, batch processing, and real-time data ingestion pipeline to move data in and out of Hadoop using PySpark and shell scripting.
  • Developed and maintained ETL/ELT processes using Informatica and debt to improve developer productivity, agility, and code quality throughout the lifecycle of data
  • Designed and implemented data models on the PostgreSQL database, including tables, views, and functions

Confidential, MA

Data Analyst / Business Analyst

Responsibilities:

  • Carried out data modeling using ER design, utilized PL/SQL objects, loaded data using ETL/Informatica
  • Developed a web application for users to upload files and get notified two weeks before data removal to avoid lengthy and troublesome data restoration processes.
  • Added modifications based on client feedback using
  • Python scripting and won an award of appreciation.
  • Good experience in AWS Services like EC2, EMR, S3, Service Catalog, and Cloud Watch.
  • Python scripts were written to read CSV, JSON, and parquet files from S3 buckets and load them into AWS S3, Dynamo DB, and Snowflake.
  • Worked on building the test cases for the Lambda function by mocking different services.
  • Responsible for writing SQL Queries and Procedures using DB2.
  • Involved in the migration of Teradata queries into the Snowflake Data warehouse queries.
  • Coding complex Oracle stored procedures, functions, packages, and cursors for client-specific applications.
  • Good experience in handling data manipulation using Python Scripts.
  • CI/CD using Git, Docker, Jenkins, and version control tools like Stash
  • Participated in database design, development, implementation, and methodologies for OTAP and OLAP database systems with team members based on customer requirements.

Confidential

Data Analyst / Business Analyst

Responsibilities:

  • Developed and modified Sybase packages, wrote and optimized in-application SQL, PL/SQL statements
  • Created DB2 database objects such as schema, tables, indexes, constraints, functions, stored procedures, views, and triggers, and maintained relationships between tables
  • Load data into Amazon Redshift and use AWS Cloud Watch to collect and monitor AWS RDS instances within Confidential.
  • Implemented AWS Step Functions to automate and orchestrate the Amazon Sage Maker-related tasks such as publishing data to S3, training the ML model, and deploying it for prediction.
  • Database Application development using Oracle PLSQL that includes Cursors, Ref-cursors, Bulking techniques,
  • Procedures, Functions, Packages, Oracle Supplied Packages, Collections, Partitioned Tables, Triggers, Materialized Views, and Table Indexing.
  • Environment: AWS EMR, S3, RDS, Redshift, Lambda, Boto3, Dynamo DB, Amazon Sage Maker
  • Apache Spark, HBase, Apache Kafka, HIVE, SQOOP, Map Reduce, Snowflake, Apache Pig, Python, SSRS, Tableau
  • Experience in using Spark SQL to handle structured data from Hive in the AWS EMR Platform
  • Implemented Job Monitoring, health checks, and job streaming of daily user production data volumes in the HDFS system using Scala, Python, Spark SQL, HBase, and Phoenix.
  • SSIS and SSRS report developments in Microsoft SQL Server.
  • Create/Modify shell scripts for scheduling various data cleansing scripts and ETL loading processes.
  • Tuned and optimized the performance of queries with modification in DB2 queries, T-SQL queries, normalized tables, established joins, and created indexes wherever necessary

Confidential

Data Analyst

Responsibilities:

  • Created database and database objects such as tables, indexes, constraints, stored procedures, views, and triggers, and maintained relationships between tables
  • Transformed the data using AWS Glue dynamic frames with PySpark; cataloged the transformed data using Crawlers and scheduled the job and crawler using workflow feature
  • Allotted permissions, policies, and roles to users and groups using AWS Identity and Access Management (IAM).
  • Redesign, and remodel the silver layer database to move Data Warehouse to the gold layer.
  • Convert looker ETL process to manage in Redshift
  • Analyse, improve and manage ETL architecture and provide strategy related to data cleansing, data quality, and data consolidation.
  • Executed the program by using Python API written in Python to support Apache Spark or PySpark.
  • Developed Informatica mappings to load data files sourced from
  • Project plan, requirement analysis, low-level design specification, development, preparation of test plans and test specs for QA/UAT, Preparation of release notes
  • Code Reviews for the team members.
  • Propose ideas to improve the process efficiency and quality through weekly and monthly status reports/meetings.

We'd love your feedback!