We provide IT Staff Augmentation Services!

Sr. Hadoop Consultant/architect Resume

4.00/5 (Submit Your Rating)

Pleasanton, CA

SUMMARY:

  • 17+ years of experience in the Information Technology industry and working in a Global context.
  • Hands - on technology leader in BI, Java and Oracle technologies with exposure to all levels of software development
  • Worked extensively on Business Intelligence / Analytics solutions.
  • Good knowledge and experience in developing company processes for code development, quality assurance and deployment.
  • Experience building BigData solutions using MapReduce, Hive, Pig, Flume, Sqoop, Oozie, Spark on Company private clouds and over AWS.
  • Hands-on experience developing Java MapReduce programs, Hive Schema design and Development, Pig script development for data ingestion and data import/export of SQOOP.
  • Working knowledge of NoSQL dbs like HBase and MongoDB.
  • Hands-on experience building data pipelines using Cascading framework.
  • Experienced building search application using Apache Solr.
  • Experienced designing and Developing DW/BI solutions using Informatica (ETL), Terdata and BOXI and Cognos.
  • Strong presentation and communication skills with the ability to mentor and motivate teams.
  • Excellent team player and coach for junior team members.
  • Experience conducting Training to Clients across Asia, Americas and Europe.
  • Contributed a couple of Hive utilities to the GitHub community
  • Scrum Master and a PMP Certified Individual. Managed BI solution development for a business acquisition project.
  • Experience working in a Global setup with virtual teams.

TECHNICAL SKILLS:

OperatingSystems: Unix, Linux, Windows NT/XP

Languages: Java, C, C++, OOAD, Python, PLSQL, XML

IDEs/Build Tools: Eclipse IDE, Version Control (SVN, Git, GitHub), Build (Maven,), Team City (CI), Kraken (Deployment).

Frameworks: Cascading, CrunchDatabases (SQL/NoSQL) Oracle 11g, Teradata

Databases (SQL/NoSQL): HBase, Mongo-DB, Solr 4.4BigData Tools/Platforms CDH 4.3.1/5.1 with Kerboros, Sentry, Hue, Hive, Impala, Pig, MR (using Java), SQOOP, Flume, Cloudera Search, Apache Solr, Morphline Framework

BI Tools/Platforms: Informatica 9.1, BO, CognosProject Management MS Office, Rally (Sprint), JIRA

CAREER PROGRESSION:

Sr. Hadoop Consultant/Architect

Confidential, Pleasanton, CA

Responsibilities:

  • Design the dataflows from ingestion to processing and display in Presentation Layer.
  • Framework development for loading and processing approx 50 to 70 TB of data.
  • Lead a team of Senior developers to have the solution developed and deployed
  • Interface point with all the functional teams Release team, infrastructure team, business teams etc.
  • Explore newer tools and technologies to help business gain more productivity. Solr features, QPL support, Cloudera 5.1 to 5.4 upgrade, Java MR to Cascading framework, Hadoop HDFS to HBase etc.
  • Design and development of rule engine which is an in-memory solution to filter fraudulant transactions.
  • Develop MR jobs, Hive scripts, Sqoop scripts which are required for job completion.
  • As an Additional responsibility, help the Confidential management turn the project into a Scrum style of development. Successfully implemented in Phase 2.0 of the project.

Technology: Java 1.7, CDH 4.6.1/5.1/5.3.1/5.4 , Sqoop (oraOop), Hive, MapReduce v1, Hbase, Oozie, Oracle, Flume, Solr 4.1/4.4/4.10 , Morphline

Sr. Hadoop Engineer - Data Engineering

Confidential, Portland, OR

Responsibilities:

  • Develop/Maintain Avro parser API (Java Spring Framework) to help build the product level dataset.
  • Develop MapReduce jobs to transform the raw sports data into product dataset using AvroInputFormat and AvroSerde in Hive tables. Approximately 20TB of data for 2013 ingested for processing.
  • Develop FLUME stream for twitter data for a POC purpose to ingest Twitter data.
  • Work with the Analysts to build the datasets for testing and training their Models .
  • Work with all the dependent teams (source teams, project management, Hadoop Admin) to resolve issues for the users.
  • Develop UDFs in Hive for custom function development (geodistance, deviceType, etc. ).
  • Development of Shell Scripts to build a workflow as mandated by the organization.
  • Development of SQOOP jobs to ingest approx. 200GB initial and 20GB per day.
  • Development of Datapipeline using Cascading Framework to replace existing MapReduce Jobs.
  • Contributor to GitHub(venrad) and StackOverflow (venBigData)

Technology: Java 1.7, CDH4.5, Sqoop, Hive, MapReduce v1, Avro serialization,, Oracle, Flume. Cascading Framework, Shell Scripts, PigBigData Consultant

Confidential, CA

Responsibilities:

  • Develop Training plan and content for the Hadoop Training
  • Interview and select candidates for the training
  • Maintain AWS and S3 environment used for training. which includes installation of hadoop and echo system tools, stop and start instances and clusters, Configure cluster when nodes added or removed.
  • Train on Hive, Pig, MapReduce Framework, Flume, Sqoop and Oozie.

Technology: Java 1.7, Apache Hadoop 1.2, Sqoop, Hive, MapReduce v1, Oracle PL/SQL, Flume. Shell Scripts, Pig, Amazon AWS /S3.

BI Technical Leader

Confidential

Responsibilities:

  • Own Development and Architecture of the entire BI piece of the IT Integration program
  • Client Management and Vendor Management.
  • Scrum Master and Project Manager for the entire program.
  • Project Management and client Management.
  • Budget ownership to drive this approx $2 million across Europe and APAC.
  • Support the Project management in the US on the US deployments.
  • Requirements development and Scope Management.

We'd love your feedback!