We provide IT Staff Augmentation Services!

Big Data Engineer Resume

2.00/5 (Submit Your Rating)

Alpharetta, GA

SUMMARY

  • Self - motivated and highly enthusiastic IT professional having 13+ years of experience in Software Development including requirement analysis, design, coding, testing, debugging and maintenance. Having perfect blend of development, data analysis, descriptive statistics and data engineering from extract, load to transform using various data mining tools makes a true profile for all types of Big Data projects. Domain experience includes, Telecom, Analytics (Nielsen Media), Insurance.
  • Well versed in all stages of Software Development Life Cycle including requirement analysis, design, development, testing, deployment and support and maintenance with various methodologies such as Waterfall and Agile models including Scrum.
  • Extensive knowledge and experience in Big data technologies Confidential research level using various tools to achieve results that are calculated on revenue and expenditure savings.
  • Well versed with Python Pandas and R to develop data driven insights.
  • Hands on experience on various RDBMS and data warehouse. Extracted, loaded and transformed data from SQL servers, Teradata, Postgres, Oracle, Sybase, DB2. Highly knowledgeable on Azure Databricks.
  • Highly experienced in creating data Pipelines and BRD (Business Ready Dataset).

TECHNICAL SKILLS

Programming Languages and Tools: Linux Bash/Shell Script, Perl script, Python (Pandas, selenium, numpy), R, Tableau, MicroStrategy, S-Lang, GitHub, Jira, Core Java, Jenkins, MS Excel, MS Access, MS PowerPoint, TOAD

Data Bases/Warehouse: Oracle, Sybase, DB2, SQL Server 2008, MySQL, Teradata, Postgres

Big Data: HDFS (Hadoop Distributed File System), Pig, Hive, HBase, Flume, Sqoop, R, MongoDB, Pyspark, NoSQL, Azure cloud, Azure Databricks

PROFESSIONAL EXPERIENCE

Confidential - Alpharetta, GA

Big Data Engineer

Responsibilities:

  • Performed data analytics using big data solutions for generating data driven insights to save revenue.
  • Descriptive statistics using R on Rcloud, Python (Pandas, numpy etc.,) for identifying and improving revenue leakages, reuse of various network equipment. Helped Confidential &T save $1B an year in capital expenditure and $1M in operational expenditure.
  • Extensively worked with data engineers and data stewards to understand data requirements for data engineering, transformation, processing. Implemented data validation checks to ensure data transfers were in acceptable ranges.
  • Worked with business partners to understand requirements and bring in right analytics to the desk. Tableau and MicroStrategy where extensively used for analysis and visualization.
  • Web scarping using python selenium was implemented to scrap data from websites.
  • Worked with legal and data governance & compliance teams to secure customer personal information & sensitive data is not exposed during any phase of analytics.
  • Analyze and arrive Confidential solutions using different data sources & structures for identifying improvements within the organization that would help in better services to customers, improve/optimize network, cost reduction in various fields and provide innovative business strategies to Confidential &T clients.

Confidential

IT Analyst

Responsibilities:

  • Hadoop was setup to help market and design better insurance policies for their existing customers.
  • Used Sqoop to transfer data between databases and HDFS. Used Flume to stream the log data from social media sources.
  • For the data loaded in HDFS, customer’s public Insurance comments/tweets from social networking sites (Twitter and Facebook) are retrieved using Flume. Data analysis was performed using Hive Serdes.

Confidential, OH

IT Analyst

Responsibilities:

  • Understanding banking and telecom client’s business and provide solutions within Hadoop eco systems. I worked on Telecom and banking e-Statement POC’s.
  • POC on market analysis/predictive analysis for telecom operator based on existing user orders data for past 10 years, which would help telecom customer to plan and increase their business. Hive was used to analyze peak and off-peak hours of orders placed, top 10 customers who placed most orders, number of disconnected orders. e-Statement generation processing time reduction for banking customer using Pig, Hive and Sqoop. POC resulted 94% processing time reduction.
  • Twitter Sentiment Analysis: Flume is used for data collection from Twitter and Hive is used to analyze data. As Twitter returns tweets in JSON format, hive serdes java library helps understand JSON format.

Confidential, Tampa, FL

IT Analyst

Responsibilities:

  • Development using UNIX shell, Perl in extending the business for Advertisement viewership analysis across country.
  • Coordinating with offshore team for ASP.Net based web application, interfacing business users in requirement analysis, design and deliver of work packets.
  • Setup of History server to provide historical data access to business users and archiving data every year.
  • Migrated and deployed code through different environments (Development, Testing, Production, and staging) and handled all kinds of change management requirements by interacting with support teams, DBA s, and System Administrators.
  • Requirements gathering from Business Users and converting them into system requirement specifications. Detailing the technical requirements for front end application.

Confidential

IT Analyst

Responsibilities:

  • Worked for Verizon FiOs system being part of TCS which provides IT services.
  • Automations using Unix shell & SQL to spool data from DB2 Database
  • Development using Unix Shell, Perl and C++
  • Production support to the DBA team

We'd love your feedback!