We provide IT Staff Augmentation Services!

Etl-hadoop Analyst Resume

4.00/5 (Submit Your Rating)

Irving, TX

PROFESSIONAL SUMMARY:

  • 8+ Years of IT experience as Developer on Mainframes Programming, Configuration Engineer on Endevor Configuration Management, Unix Prod Support in AIX & Solaris, Development & Enhancement of Data warehouses technologies And Hadoop Development.
  • Had 3 year and 3 months of IT experience of complete lifecycle implementations from design and migration of the elements as a Configuration Engineer through ENDEVOR TOOL and Unix Prod Support.
  • Had In - depth Expertise in Telecom Domain.
  • Had Comprehensive Experience of 1 year and 5 months of IT experience in Development & Enhancement of Data warehouses and have worked extensively on Informatica 9.5.1, UNIX, Tidal and Teradata 14.
  • Had Experience of 3 years and 4 months as an Apache Hadoop Developer. Expertise in writing Hadoop Jobs for analyzing data using HDFS, Hive, HBase, Pig, Spark, Scala and Gaining knowledge on Oozie
  • Good knowledge of Hadoop Architecture and various components such as HDFS, Job Tracker, Task Tracker, Name Node, Data Node and MapReduce concepts.
  • Working experience on designing and implementing complete end-to-end Hadoop Infrastructure including PIG, HIVE, Sqoop, Flume/ Kafka.
  • Performed Sales to Billing Analysis on Video Applications and Churn Analysis using Data on Hadoop Cluster.
  • Had Good knowledge in writing MAP REDUCE Programs
  • Excellent communication, demonstrated interpersonal and leadership skills
  • Good Knowledge of the Configuration Management tool -Endeavor,
  • Good knowledge on Solaris 10, IBM AIX 5.3L, Backup Administration
  • Good Knowledge on DB2-DBA and DB2-UDB admin concepts.

TECHNICAL SKILLS:

BIG DATA: Apache Hadoop 2.6.5, CloudEra Enterprise 4.X

NO SQL databases: HBASE 0.94.5

Hadoop Ecosystems: HIVE 0.10.0, PIG 0.11.0, Sqoop 1.4.3, Flume 1.4.0 and KAFKA

Operating Systems: Z/OS, Solaris 10, IBM AIX 5.3L, Ubuntu 13.X Languages SCALA, COBOL,SQL and Core java and Python

Databases: DB2-DBA, DB2-UDB, Teradata 14, HIVE

Configuration TOOLS: ENDEVOR

ETL TOOLS: Informatica 9.5.1

Applications: MS office

Development Tools: SSH Secure Shell, Teradata SQL Assistance

Scheduling Tools: Tidal, ESP, CRON TAB on UNIX

PROFESSIONAL EXPERIENCE:

Confidential, Irving TX

ETL-Hadoop Analyst

Role/Responsibilities:

  • Identify the Customer Usage, Behavior, Feedback datasets which would reside on different systems and unfolding insights in to Customer Usage, behavior and Feedback analysis who are likely to Churn.
  • Analyzed data which need to be loaded into hadoop and contacted with respective source teams to get the table information and connection details.
  • Migrating the data of high volumes from Oracle, MySQL in to HDFS using Sqoop, Informatica ETL and importing various formats of flat files in to HDFS.
  • Creating Partitioned tables HIVE. Mainly worked on Hive QL to categorize data of different claims, Implemented Partitioning, Dynamic Partitions, Buckets in HIVE.
  • Analyzed large amounts of data sets to determine optimal way to aggregate and report on it.
  • Working on projecting, involving and migration of data from the mainframes, Teradata to HDFS Data Lake and creating reports by performing transformations on the data put in the Hadoop data lake.
  • Cleansing the Data and developing a strategy for Full load and incremental load using Sqoop.
  • Analyze data in Pig Latin, Hive and Map Reduce in Java and SCALA (SPARK).
  • Import tables from MySQL database to HDFS and HBase using Sqoop.
  • Had knowledge on workflows for Map Reduce, Pig and Hive jobs in Oozie.
  • Gained knowledge in integrating the hive warehouse with Hbase.
  • Constantly worked on tuning the performance of the queries in Hive and Pig, making the queries work even more powerfully in processing and retrieving the data
  • Along with the Infrastructure team, involved in design and developed Kafka and Streaming data pipeline
  • Used RDD's to perform transformation on datasets as well as to perform actions like count, reduce, first.
  • Documented ETL best practices to be implemented with Hadoop
  • Monitoring and Debugging Hadoop jobs/Applications running in production.
  • As a team member monitored and supported on Hadoop Cloudera upgrade from CDH3 to CDH4.x.
  • Develop HIVE and PIG Jobs and Worked with Data-scientists on Sentimental Analysis to identify the locations where percentage of CHURN is high and identify customers performing rotational Churns
  • Preparing the Design, Approach and Solution documents and Follow the SDLC cycle for production implementations.

Confidential

Sr. Informatica developer

Role/Responsibilities:

  • Extracting, transforming and loading data from Various sources on Daily, weekly and monthly basis into the database with the help of batch jobs to ensure the data loaded is 100 % fine.
  • Supported Developed mappings & updating the existing mappings as per the requirement.
  • Analysis of existing code and testing of the programs as per the requirement.
  • Monitoring the Incident Management queue & fixing the issues with appropriate resolution steps and maintaining proper documentation for further reference.
  • Comparing the record counts with the source file counts by using Informatica monitor.
  • Involved creating tickets, creating Change Activity documents etc., when the New Changes in code has to promote to production.
  • Used most of the transformations such as the Source Qualifier, Expression, Aggregator, Connected & unconnected lookups, Filter, Router, Sequence Generator, Sorter, Joiner, and Update Strategy.
  • Imported data from various Sources transformed and loaded into Data Warehouse Targets using Informatica.
  • Extensively worked on data extraction, transformation and data loading from source to target system using Teradata Utilities BTEQ and Fastload, MLOAD according to Business needs.
  • Documentation (LLD) to describe program development, logic, coding, testing, changes and corrections.
  • Interacting with business users to let them know about the availability of data and imbalanced data.
  • Monitor the Daily, Weekly and monthly production loads in Tidal Scheduler.

Confidential

Technical Staff Member

Role/Responsibilities:

  • Attending e-mail queries from the users, alerts on CMAT Verizon Proprietary web tool
  • Executing Code Cut to Dev Environment, Scheduled and Immediate distributions, roll the code in test Environments and Upgrading Production environments.
  • Working on Different requests. Archiving the Prod Code Quarterly and Regular release.
  • Maintaining 24x7 CM control desk with Endevor and ESP scheduling tool
  • Attending client meetings, turnover meetings and status calls and Coordinating the configuration management activities with the offshore/Onshore team.
  • Reviewing and providing feedback to the work done by offshore.
  • Interacting with users and help them fix any Endevor or Configuration issues.

Confidential

Software Engineer

Role/Responsibilities

  • Working on Create and publish the Master Element List (MEL). Notify Release Team of list availability
  • Work involved and to Ensure Development has Retrieved, w/Sign Out, all the elements on the Master Element List with the correct CCID.
  • Doing the Pre-Implementation activities and involving in the Implementation activities and Work on Delete requests on customer request.
  • Attending the Release Kick off meetings, working along with the release schedule and following up the development teams for proper sign into the Release Paths.
  • System installation which includes new & complete overwrites migration and configuring the devices & handling the hardware problems.
  • Using logical volume Manager to manage VG’s PV’s & LV’s. Creating, deleting & changing the characteristics of PV’s, LV’s & VG’s. Exporting & Importing VG’s.
  • Managing the file systems, and Monitoring the performance of system
  • Adding, removing the user accounts, changing the characteristics of users like login shell, shell prompt, password etc.
  • Backing up the system using mksysb, backing up user information, restoring information from backup media..

We'd love your feedback!