We provide IT Staff Augmentation Services!

Sr. Hadoop Consultant/architect Resume

2.00/5 (Submit Your Rating)

Pleasanton, CA

SUMMARY

  • 17+ years of experience in teh Information Technology industry and working in a Global context.
  • Hands - on technology leader in BI, Java and Oracle technologies wif exposure to all levels of software development
  • Worked extensively on Business Intelligence / Analytics solutions.
  • Good knowledge and experience in developing company processes for code development, quality assurance and deployment.
  • Experience building BigData solutions using MapReduce, Hive, Pig, Flume, Sqoop, Oozie, Spark on Company private clouds and over AWS.
  • Hands-on experience developing Java MapReduce programs, Hive Schema design and Development, Pig script development for data ingestion and data import/export of SQOOP.
  • Working knowledge of NoSQL dbs like HBase and MongoDB.
  • Hands-on experience building data pipelines using Cascading framework.
  • Experienced building search application using Apache Solr.
  • Experienced designing and Developing DW/BI solutions using Informatica (ETL), Terdata and BOXI and Cognos.
  • Strong presentation and communication skills wif teh ability to mentor and motivate teams.
  • Excellent team player and coach for junior team members.
  • Experience conducting Training to Clients across Asia, Americas and Europe.
  • Contributed a couple of Hive utilities to teh GitHub community
  • Scrum Master and a PMP Certified Individual. Managed BI solution development for a business acquisition project.
  • Experience working in a Global setup wif virtual teams.

TECHNICAL SKILLS

Operating Systems: Unix, Linux, Windows NT/XP

Languages: Java, C, C++, OOAD, Python, PLSQL, XML

IDEs/Build Tools: Eclipse IDE, Version Control (SVN, Git, GitHub), Build (Maven,), Team City (CI), Kraken (Deployment).

Frameworks: Cascading, Crunch

Databases (SQL/NoSQL): Oracle 11g, TeradataDatabases (SQL/NoSQL) HBase, Mongo-DB, Solr 4.4

BigData Tools/Platforms: CDH 4.3.1/5.1 wif Kerboros, Sentry, Hue, Hive, Impala, Pig, MR (using Java), SQOOP, Flume, Cloudera Search, Apache Solr, Morphline Framework

BI Tools/Platforms: Informatica 9.1, BO, Cognos

Project Management: MS Office, Rally (Sprint), JIRA

PROFESSIONAL EXPERIENCE

Sr. Hadoop Consultant/Architect

Confidential, Pleasanton, CA

Responsibilities:

  • Design teh dataflows from ingestion to processing and display in Presentation Layer.
  • Framework development for loading and processing approx 50 to 70 TB of data.
  • Lead a team of Senior developers to has teh solution developed and deployed
  • Interface point wif all teh functional teams Release team, infrastructure team, business teams etc.
  • Explore newer tools and technologies to help business gain more productivity. Solr features, QPL support, Cloudera 5.1 to 5.4 upgrade, Java MR to Cascading framework, Hadoop HDFS to HBase etc.
  • Design and development of rule engine which is an in-memory solution to filter fraudulant transactions.
  • Develop MR jobs, Hive scripts, Sqoop scripts which are required for job completion.
  • As an Additional responsibility, help teh Confidential management turn teh project into a Scrum style of development. Successfully implemented in Phase 2.0 of teh project.

Sr. Hadoop Engineer

Confidential, Portland, OR

Responsibilities:

  • Develop/Maintain Avro parser API (Java Spring Framework) to help build teh product level dataset.
  • Develop MapReduce jobs to transform teh raw sports data into product dataset using AvroInputFormat and AvroSerde in Hive tables. Approximately 20TB of data for 2013 ingested for processing.
  • Develop FLUME stream for twitter data for a POC purpose to ingest Twitter data.
  • Work wif teh Analysts to build teh datasets for testing and training their Models .
  • Work wif all teh dependent teams (source teams, project management, Hadoop Admin) to resolve issues for teh users.
  • Develop UDFs in Hive for custom function development (geodistance, deviceType, etc. ).
  • Development of Shell Scripts to build a workflow as mandated by teh organization.
  • Development of SQOOP jobs to ingest approx. 200GB initial and 20GB per day.
  • Development of Datapipeline using Cascading Framework to replace existing MapReduce Jobs.
  • Contributor to GitHub(venrad) and StackOverflow (venBigData)

BigData Consultant

Confidential, CA

Responsibilities:

  • Develop Training plan and content for teh Hadoop Training
  • Interview and select candidates for teh training
  • Maintain AWS and S3 environment used for training. which includes installation of hadoop and echo system tools, stop and start instances and clusters, Configure cluster when nodes added or removed.
  • Train on Hive, Pig, MapReduce Framework, Flume, Sqoop and Oozie.

BI Technical Leader

Confidential

Responsibilities:

  • Own Development and Architecture of teh entire BI piece of teh IT Integration program
  • Client Management and Vendor Management.
  • Scrum Master and Project Manager for teh entire program.
  • Own a budget of approx $1 million.

BI Technical Leader

Confidential

Responsibilities:

  • Project Management and client Management.
  • Budget ownership to drive dis approx $2 million across Europe and APAC.
  • Support teh Project management in teh US on teh US deployments.
  • Requirements development and Scope Management.

We'd love your feedback!