Sr. Hadoop Consultant/architect Resume
Pleasanton, CA
SUMMARY
- 17+ years of experience in teh Information Technology industry and working in a Global context.
- Hands - on technology leader in BI, Java and Oracle technologies wif exposure to all levels of software development
- Worked extensively on Business Intelligence / Analytics solutions.
- Good knowledge and experience in developing company processes for code development, quality assurance and deployment.
- Experience building BigData solutions using MapReduce, Hive, Pig, Flume, Sqoop, Oozie, Spark on Company private clouds and over AWS.
- Hands-on experience developing Java MapReduce programs, Hive Schema design and Development, Pig script development for data ingestion and data import/export of SQOOP.
- Working knowledge of NoSQL dbs like HBase and MongoDB.
- Hands-on experience building data pipelines using Cascading framework.
- Experienced building search application using Apache Solr.
- Experienced designing and Developing DW/BI solutions using Informatica (ETL), Terdata and BOXI and Cognos.
- Strong presentation and communication skills wif teh ability to mentor and motivate teams.
- Excellent team player and coach for junior team members.
- Experience conducting Training to Clients across Asia, Americas and Europe.
- Contributed a couple of Hive utilities to teh GitHub community
- Scrum Master and a PMP Certified Individual. Managed BI solution development for a business acquisition project.
- Experience working in a Global setup wif virtual teams.
TECHNICAL SKILLS
Operating Systems: Unix, Linux, Windows NT/XP
Languages: Java, C, C++, OOAD, Python, PLSQL, XML
IDEs/Build Tools: Eclipse IDE, Version Control (SVN, Git, GitHub), Build (Maven,), Team City (CI), Kraken (Deployment).
Frameworks: Cascading, Crunch
Databases (SQL/NoSQL): Oracle 11g, TeradataDatabases (SQL/NoSQL) HBase, Mongo-DB, Solr 4.4
BigData Tools/Platforms: CDH 4.3.1/5.1 wif Kerboros, Sentry, Hue, Hive, Impala, Pig, MR (using Java), SQOOP, Flume, Cloudera Search, Apache Solr, Morphline Framework
BI Tools/Platforms: Informatica 9.1, BO, Cognos
Project Management: MS Office, Rally (Sprint), JIRA
PROFESSIONAL EXPERIENCE
Sr. Hadoop Consultant/Architect
Confidential, Pleasanton, CA
Responsibilities:
- Design teh dataflows from ingestion to processing and display in Presentation Layer.
- Framework development for loading and processing approx 50 to 70 TB of data.
- Lead a team of Senior developers to has teh solution developed and deployed
- Interface point wif all teh functional teams Release team, infrastructure team, business teams etc.
- Explore newer tools and technologies to help business gain more productivity. Solr features, QPL support, Cloudera 5.1 to 5.4 upgrade, Java MR to Cascading framework, Hadoop HDFS to HBase etc.
- Design and development of rule engine which is an in-memory solution to filter fraudulant transactions.
- Develop MR jobs, Hive scripts, Sqoop scripts which are required for job completion.
- As an Additional responsibility, help teh Confidential management turn teh project into a Scrum style of development. Successfully implemented in Phase 2.0 of teh project.
Sr. Hadoop Engineer
Confidential, Portland, OR
Responsibilities:
- Develop/Maintain Avro parser API (Java Spring Framework) to help build teh product level dataset.
- Develop MapReduce jobs to transform teh raw sports data into product dataset using AvroInputFormat and AvroSerde in Hive tables. Approximately 20TB of data for 2013 ingested for processing.
- Develop FLUME stream for twitter data for a POC purpose to ingest Twitter data.
- Work wif teh Analysts to build teh datasets for testing and training their Models .
- Work wif all teh dependent teams (source teams, project management, Hadoop Admin) to resolve issues for teh users.
- Develop UDFs in Hive for custom function development (geodistance, deviceType, etc. ).
- Development of Shell Scripts to build a workflow as mandated by teh organization.
- Development of SQOOP jobs to ingest approx. 200GB initial and 20GB per day.
- Development of Datapipeline using Cascading Framework to replace existing MapReduce Jobs.
- Contributor to GitHub(venrad) and StackOverflow (venBigData)
BigData Consultant
Confidential, CA
Responsibilities:
- Develop Training plan and content for teh Hadoop Training
- Interview and select candidates for teh training
- Maintain AWS and S3 environment used for training. which includes installation of hadoop and echo system tools, stop and start instances and clusters, Configure cluster when nodes added or removed.
- Train on Hive, Pig, MapReduce Framework, Flume, Sqoop and Oozie.
BI Technical Leader
Confidential
Responsibilities:
- Own Development and Architecture of teh entire BI piece of teh IT Integration program
- Client Management and Vendor Management.
- Scrum Master and Project Manager for teh entire program.
- Own a budget of approx $1 million.
BI Technical Leader
Confidential
Responsibilities:
- Project Management and client Management.
- Budget ownership to drive dis approx $2 million across Europe and APAC.
- Support teh Project management in teh US on teh US deployments.
- Requirements development and Scope Management.